semi supervised learning for neural machine translation
play

Semi-supervised Learning for Neural Machine Translation Yong Cheng - PowerPoint PPT Presentation

Semi-supervised Learning for Neural Machine Translation Yong Cheng joint work with Wei Xu, Zhongjun He, Wei He, Hua Wu, Maosong Sun, Yang Liu 1 Machine Translation Automated translation using computer software 2 Machine Translation Rule-based


  1. Semi-supervised Learning for Neural Machine Translation Yong Cheng joint work with Wei Xu, Zhongjun He, Wei He, Hua Wu, Maosong Sun, Yang Liu 1

  2. Machine Translation Automated translation using computer software 2

  3. Machine Translation Rule-based Machine Translation 1970s Example-based Machine Translation 1984 Statical Machine Translation (SMT) 1993 Neural Machine Translation � NMT � 2014 Trends: learning to translate from DATA 3

  4. Machine Translation Parallel corpora are usually limited in & & quantity quality coverage Monolingual Corpora Parallel Corpora 4

  5. Monolingual Corpora Used in SMT and NMT N-gram language model in SMT Koehn et al., [2007] Monolingual corpora as decipherment Ravi and Knight [2011] Integrate a neural language model into NMT. Gulccehre et al. [2015] Additional pseudo parallel corpus. Sennrich et al. [2016] 5

  6. Supervised Training Parallel Corpus Objective 6

  7. Unsupervised Training Monolingual Corpus 7

  8. cc Our Approach — Autoencoders x bushi yu shalong juxing le huitan 8

  9. cc Our Approach — Autoencoders ! θ ) P ( y | x ; x bushi yu shalong juxing le huitan 9

  10. cc Our Approach — Autoencoders y latent Bush held a talk with sharon ! θ ) P ( y | x ; x bushi yu shalong juxing le huitan 10

  11. cc Our Approach — Autoencoders ! θ ) P ( x | y ; y latent Bush held a talk with sharon ! θ ) P ( y | x ; x bushi yu shalong juxing le huitan 11

  12. cc Our Approach — Autoencoders ′ x bushi yu shalong juxing le huitan ! θ ) P ( x | y ; y latent Bush held a talk with sharon ! θ ) P ( y | x ; x bushi yu shalong juxing le huitan 12

  13. cc Our Approach — Autoencoders source autoencoder target autoencoder 13

  14. Unsupervised Training (Autoencoders) Monolingual Corpus target autoencoder 14

  15. Semi-supervised Training Training Objective 15

  16. Translation Results Compared with Moses (SMT) and RNNSearch (NMT) 16

  17. Translation Results Compared with Moses (SMT) and RNNSearch (NMT) 17

  18. Translation Results Compared with Moses (SMT) and RNNSearch (NMT) 18

  19. Translation Results Compared with Moses (SMT) and RNNSearch (NMT) 19

  20. Translation Results Compared with Moses (SMT) and RNNSearch (NMT) 20

  21. Translation Results Compared with Sennrich et al. [2015a] 21

  22. Example Translation of Monolingual Corpus 22

  23. Conclusion Monolingual corpora is an important resource for neural machine translation. We have proposed a semi-supervised approach to training bidirectional neural machine translation models for exploiting monolingual corpora. As our method is sensitive to the OOVs present in monolingual corpora, we plan to integrate Jean et al. (2015)’s technique on using very large vocabulary into our approach. 23

  24. Thank You ! 24

  25. Effect of Sample Size ZH-EN EN-ZH 25

  26. Effect of OOV ratio ZH-EN EN-ZH 26

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend