A Survey On Contextual Embeddings | Awesome LLM Papers Contribute to Awesome LLM Papers

A Survey On Contextual Embeddings

Qi Liu, Matt J. Kusner, Phil Blunsom . Arxiv 2020 – 113 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
Uncategorized

Contextual embeddings, such as ELMo and BERT, move beyond global word representations like Word2Vec and achieve ground-breaking performance on a wide range of natural language processing tasks. Contextual embeddings assign each word a representation based on its context, thereby capturing uses of words across varied contexts and encoding knowledge that transfers across languages. In this survey, we review existing contextual embedding models, cross-lingual polyglot pre-training, the application of contextual embeddings in downstream tasks, model compression, and model analyses.

Similar Work