Non-autoregressive Transformer By Position Learning · Awesome LLM Papers Contribute to LLM-Bible

Non-autoregressive Transformer By Position Learning

Yu Bao et al.. Arxiv 2019 – 29 citations

[Paper]    
Transformer Language Modeling GPT Model Architecture

Non-autoregressive models are promising on various text generation tasks. Previous work hardly considers to explicitly model the positions of generated words. However, position modeling is an essential problem in non-autoregressive text generation. In this study, we propose PNAT, which incorporates positions as a latent variable into the text generative process. Experimental results show that PNAT achieves top results on machine translation and paraphrase generation tasks, outperforming several strong baselines.

Similar Work