Are Pre-trained Language Models Aware Of Phrases? Simple But Strong Baselines For Grammar Induction | Awesome LLM Papers Add your paper to Awesome LLM Papers

Are Pre-trained Language Models Aware Of Phrases? Simple But Strong Baselines For Grammar Induction

Taeuk Kim, Jihun Choi, Daniel Edmiston, Sang-Goo Lee . Arxiv 2020 – 56 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
Compositional Generalization Content Enrichment Image Text Integration Interactive Environments Interdisciplinary Approaches Multimodal Semantic Representation Neural Machine Translation Productivity Enhancement Question Answering Training Techniques

With the recent success and popularity of pre-trained language models (LMs) in natural language processing, there has been a rise in efforts to understand their inner workings. In line with such interest, we propose a novel method that assists us in investigating the extent to which pre-trained LMs capture the syntactic notion of constituency. Our method provides an effective way of extracting constituency trees from the pre-trained LMs without training. In addition, we report intriguing findings in the induced trees, including the fact that pre-trained LMs outperform other approaches in correctly demarcating adverb phrases in sentences.

Similar Work