Developing Real-time Streaming Transformer Transducer For Speech Recognition On Large-scale Dataset | Awesome LLM Papers Add your paper to Awesome LLM Papers

Developing Real-time Streaming Transformer Transducer For Speech Recognition On Large-scale Dataset

Xie Chen, Yu Wu, Zhenghao Wang, Shujie Liu, Jinyu Li . ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2021 – 149 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
Applications Datasets ICASSP Interdisciplinary Approaches Model Architecture

Recently, Transformer based end-to-end models have achieved great success in many areas including speech recognition. However, compared to LSTM models, the heavy computational cost of the Transformer during inference is a key issue to prevent their applications. In this work, we explored the potential of Transformer Transducer (T-T) models for the fist pass decoding with low latency and fast speed on a large-scale dataset. We combine the idea of Transformer-XL and chunk-wise streaming processing to design a streamable Transformer Transducer model. We demonstrate that T-T outperforms the hybrid model, RNN Transducer (RNN-T), and streamable Transformer attention-based encoder-decoder model in the streaming scenario. Furthermore, the runtime cost and latency can be optimized with a relatively small look-ahead.

Similar Work