Improving Passage Retrieval With Zero-shot Question Generation | Awesome LLM Papers Add your paper to Awesome LLM Papers

Improving Passage Retrieval With Zero-shot Question Generation

Devendra Singh Sachan, Mike Lewis, Mandar Joshi, Armen Aghajanyan, Wen-Tau Yih, Joelle Pineau, Luke Zettlemoyer . Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing 2022 – 47 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
Datasets EMNLP Interdisciplinary Approaches Question Answering Training Techniques

We propose a simple and effective re-ranking method for improving passage retrieval in open question answering. The re-ranker re-scores retrieved passages with a zero-shot question generation model, which uses a pre-trained language model to compute the probability of the input question conditioned on a retrieved passage. This approach can be applied on top of any retrieval method (e.g. neural or keyword-based), does not require any domain- or task-specific training (and therefore is expected to generalize better to data distribution shifts), and provides rich cross-attention between query and passage (i.e. it must explain every token in the question). When evaluated on a number of open-domain retrieval datasets, our re-ranker improves strong unsupervised retrieval models by 6%-18% absolute and strong supervised models by up to 12% in terms of top-20 passage retrieval accuracy. We also obtain new state-of-the-art results on full open-domain question answering by simply adding the new re-ranker to existing models with no further changes.

Similar Work