RETURNN: The RWTH Extensible Training Framework For Universal Recurrent Neural Networks | Awesome LLM Papers Add your paper to Awesome LLM Papers

RETURNN: The RWTH Extensible Training Framework For Universal Recurrent Neural Networks

Patrick Doetsch, Albert Zeyer, Paul Voigtlaender, Ilya Kulikov, Ralf SchlΓΌter, Hermann Ney . 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2017 – 71 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
Evaluation ICASSP Interdisciplinary Approaches Model Architecture Neural Machine Translation Tools Training Techniques Variational Autoencoders

In this work we release our extensible and easily configurable neural network training software. It provides a rich set of functional layers with a particular focus on efficient training of recurrent neural network topologies on multiple GPUs. The source of the software package is public and freely available for academic research purposes and can be used as a framework or as a standalone tool which supports a flexible configuration. The software allows to train state-of-the-art deep bidirectional long short-term memory (LSTM) models on both one dimensional data like speech or two dimensional data like handwritten text and was used to develop successful submission systems in several evaluation campaigns.

Similar Work