rapid adaptation of machine translation to new languages
play

Rapid Adaptation of Machine Translation to New Languages Graham - PowerPoint PPT Presentation

Rapid Adaptation of Machine Translation to New Languages Graham Neubig, Junjie Hu @ EMNLP 11/2/2018 Inspiration: Rapid Disaster Response - #HiruNews #StandBy


  1. Rapid Adaptation of Machine Translation to New Languages Graham Neubig, Junjie Hu @ EMNLP 11/2/2018

  2. Inspiration: Rapid Disaster Response නාවල��ය යටකළ වැ� ව�ර - #HiruNews #StandBy ගංව�ර හා නායයා� ත�වය�ෙග� �පතට ප� වූ ���වලට සහන සැපයූ �ෙ��ඡා ක�ඩාය� ද�ව�ට ���� ඉ�ලා පළ කළ සමාජ ජාල ප�වුඩ පසු�ය �නවල ද��නට ලැ�� . Disaster in Sri Lanka Photo Credit: Wikimedia Commons

  3. How can we effectively and rapidly adapt MT to new languages?

  4. Some Crazy Ideas • Cross-lingual transfer: can we create a machine translation system by transferring across language boundaries? [Zoph+16] • Zero-shot transfer: can we do it with no data in the low- resource language?

  5. Multi-lingual Training [Firat+16, Johnson+17, Ha+17] • Train a large multi-lingual MT system, and apply it to a low-resource language fra por rus eng tur ... bel aze

  6. Two Multilingual Training Paradigms • Warm-start training: (indicated w/ "+") fra por • We already have some data in the test language rus eng tur • Train a model starting with that data . bel • Cold-start training: (indicated w/ "-") aze fra • We initially have no data in the test language por rus • Possibilities for completely unsupervised transfer eng tur . • Suitable for rapid adaptation to new languages bel x aze

  7. Experiments: Training Setting • TED multi-lingual corpus (Qi et al. 2018) 
 https://github.com/neulab/word-embeddings-for-nmt • 57 source languages, plus English • Testbed languages: Azerbaijani (aze), Belarusian (bel), Galician (glg), Slovak (slk) • Related languages: Turkish (tur), Russian (rus), Portuguese (por), Czech (ces)

  8. Systems • Test Systems: • Single-source Neural MT (Sing.): Test source language only • Bi-source Neural MT (Bi.): Test source language and related source • All-source Neural MT (All): All source languages • Other Baselines: • Phrase-based MT: Shown to be strong in low-resource settings • Unsupervised MT [Artetxe+17]: Learn system using only monolingual data in source/target languages (cited as effective in low-resource settings)

  9. How does Cross-lingual Transfer Help? PBMT Unsupervised NMT Sing. NMT Bi+ NMT All+ 30 22.5 15 7.5 0 aze/tur bel/rus glg/por slk/ces • Unsupervised translation not competitive • Without transfer, NMT worse than PBMT • With transfer NMT significantly better (transfer barely helped PBMT)

  10. How Does Cold-start Compare? NMT Bi+ NMT All+ NMT Bi- NMT All- 30 22.5 15 7.5 0 aze/tur bel/rus glg/por slk/ces • Large drop, but still much better than nothing • Up to 15 BLEU with no training data in test language

  11. Adaptation to New Languages • Training on all languages can be less effective, esp. in cold-start case • Can we further adapt to new languages? • Problem: overfitting Adaptation (All → Sing.) Pre-training aze eng fra por Adaptation w/ rus eng Similar Language Regularization tur (All → Bi.) ... bel tur eng aze aze

  12. Warm-start + Adaptation NMT Sing. NMT Bi+ NMT All+ All+ -> Sing. All+ -> Bi 30 22.5 15 7.5 0 aze/tur bel/rus glg/por slk/ces • Adaptation helps! • Helps more w/ similar language regularization

  13. Cold-start + Adaptation NMT Sing. NMT Bi- NMT All- All- -> Sing. All- -> Bi All+ -> Bi 30 22.5 15 7.5 0 aze/tur bel/rus glg/por slk/ces • Adaptation w/ similar-language regularization gains more • Approaches quality of warm-start; doesn't need data a-priori

  14. How Fast can we Adapt? Cold-start adaptation reaches good point faster than training from scratch 0.21 0.18 0.15 Sing. 0.12 Bi BLEU 0.09 All-→Sing. All-→Bi 0.06 All-→Bi 1-1 0.03 0 0 1 2 3 4 5 6 7 8 9 10 Hours Training

  15. Take-aways • NMT with massively multi-lingual cross-lingual transfer : a stable recipe for low- resource translation • Better results than phrase-based, unsupervised MT in real low-resource languages • Adaptation w/ similar language regularization : simple and effective, even in cold- start scenarios https://github.com/neubig/rapid-adaptation Questions?

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend