From Balustrades To Pierre Vinken: Looking For Syntax In Transformer Self-attentions | Awesome LLM Papers Add your paper to Awesome LLM Papers

From Balustrades To Pierre Vinken: Looking For Syntax In Transformer Self-attentions

David Mareček, Rudolf Rosa . Proceedings of the 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP 2019 – 47 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
ACL Interdisciplinary Approaches Model Architecture Variational Autoencoders

We inspect the multi-head self-attention in Transformer NMT encoders for three source languages, looking for patterns that could have a syntactic interpretation. In many of the attention heads, we frequently find sequences of consecutive states attending to the same position, which resemble syntactic phrases. We propose a transparent deterministic method of quantifying the amount of syntactic information present in the self-attentions, based on automatically building and evaluating phrase-structure trees from the phrase-like sequences. We compare the resulting trees to existing constituency treebanks, both manually and by computing precision and recall.

Similar Work