Trivial Transfer Learning For Low-resource Neural Machine Translation | Awesome LLM Papers Contribute to Awesome LLM Papers

Trivial Transfer Learning For Low-resource Neural Machine Translation

Tom Kocmi, Ondřej Bojar . Proceedings of the Third Conference on Machine Translation: Research Papers 2018 – 158 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
Uncategorized WMT

Transfer learning has been proven as an effective technique for neural machine translation under low-resource conditions. Existing methods require a common target language, language relatedness, or specific training tricks and regimes. We present a simple transfer learning method, where we first train a “parent” model for a high-resource language pair and then continue the training on a lowresource pair only by replacing the training corpus. This “child” model performs significantly better than the baseline trained for lowresource pair only. We are the first to show this for targeting different languages, and we observe the improvements even for unrelated languages with different alphabets.

Similar Work