Massively Multilingual Neural Machine Translation In The Wild: Findings And Challenges | Awesome LLM Papers Contribute to Awesome LLM Papers

Massively Multilingual Neural Machine Translation In The Wild: Findings And Challenges

Naveen Arivazhagan, Ankur Bapna, Orhan Firat, Dmitry Lepikhin, Melvin Johnson, Maxim Krikun, Mia Xu Chen, Yuan Cao, George Foster, Colin Cherry, Wolfgang MacHerey, Zhifeng Chen, Yonghui Wu . Arxiv 2019 – 320 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
Uncategorized

We introduce our efforts towards building a universal neural machine translation (NMT) system capable of translating between any language pair. We set a milestone towards this goal by building a single massively multilingual NMT model handling 103 languages trained on over 25 billion examples. Our system demonstrates effective transfer learning ability, significantly improving translation quality of low-resource languages, while keeping high-resource language translation quality on-par with competitive bilingual baselines. We provide in-depth analysis of various aspects of model building that are crucial to achieving quality and practicality in universal NMT. While we prototype a high-quality universal translation system, our extensive empirical analysis exposes issues that need to be further addressed, and we suggest directions for future research.

Similar Work