Adventure: Adversarial Training For Textual Entailment With Knowledge-guided Examples | Awesome LLM Papers Contribute to Awesome LLM Papers

Adventure: Adversarial Training For Textual Entailment With Knowledge-guided Examples

Dongyeop Kang, Tushar Khot, Ashish Sabharwal, Eduard Hovy . Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2018 – 69 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
ACL Uncategorized

We consider the problem of learning textual entailment models with limited supervision (5K-10K training examples), and present two complementary approaches for it. First, we propose knowledge-guided adversarial example generators for incorporating large lexical resources in entailment models via only a handful of rule templates. Second, to make the entailment model - a discriminator - more robust, we propose the first GAN-style approach for training it using a natural language example generator that iteratively adjusts based on the discriminator’s performance. We demonstrate effectiveness using two entailment datasets, where the proposed methods increase accuracy by 4.7% on SciTail and by 2.8% on a 1% training sub-sample of SNLI. Notably, even a single hand-written rule, negate, improves the accuracy on the negation examples in SNLI by 6.1%.

Similar Work