Marian: Cost-effective High-quality Neural Machine Translation In C++ | Awesome LLM Papers Add your paper to Awesome LLM Papers

Marian: Cost-effective High-quality Neural Machine Translation In C++

Marcin Junczys-Dowmunt, Kenneth Heafield, Hieu Hoang, Roman Grundkiewicz, Anthony Aue . Proceedings of the 2nd Workshop on Neural Machine Translation and Generation 2018 – 54 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
Interdisciplinary Approaches Model Architecture Neural Machine Translation Training Techniques

This paper describes the submissions of the “Marian” team to the WNMT 2018 shared task. We investigate combinations of teacher-student training, low-precision matrix products, auto-tuning and other methods to optimize the Transformer model on GPU and CPU. By further integrating these methods with the new averaging attention networks, a recently introduced faster Transformer variant, we create a number of high-quality, high-performance models on the GPU and CPU, dominating the Pareto frontier for this shared task.

Similar Work