TR-BERT: Dynamic Token Reduction For Accelerating BERT Inference · Awesome LLM Papers Contribute to LLM-Bible

TR-BERT: Dynamic Token Reduction For Accelerating BERT Inference

Deming Ye, Yankai Lin, Yufei Huang, Maosong Sun. Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies 2021 – 17 citations

[Paper] [Code]    
Has Code Model Architecture Attention Mechanism Transformer Applications Reinforcement Learning Agentic BERT

Existing pre-trained language models (PLMs) are often computationally expensive in inference, making them impractical in various resource-limited real-world applications. To address this issue, we propose a dynamic token reduction approach to accelerate PLMs’ inference, named TR-BERT, which could flexibly adapt the layer number of each token in inference to avoid redundant calculation. Specially, TR-BERT formulates the token reduction process as a multi-step token selection problem and automatically learns the selection strategy via reinforcement learning. The experimental results on several downstream NLP tasks show that TR-BERT is able to speed up BERT by 2-5 times to satisfy various performance demands. Moreover, TR-BERT can also achieve better performance with less computation in a suite of long-text tasks since its token-level layer number adaption greatly accelerates the self-attention operation in PLMs. The source code and experiment details of this paper can be obtained from https://github.com/thunlp/TR-BERT.

Similar Work