Hybrid Autoregressive Solver for Scalable Abductive Natural Language Inference

Regenerating natural language explanations for science questions is a challenging task for evaluating complex multi-hop and abductive inference capabilities. In this setting, Transformers trained on human-annotated explanations achieve state-of-the-art performance when adopted as cross-encoder architectures. However, while much attention has been devoted to the quality of the constructed explanations, the problem of performing abductive inference at scale is still under-studied. As intrinsically not scalable, the cross-encoder architectural paradigm is not suitable for efficient multi-hop inference on massive facts banks. To maximise both accuracy and inference time, we propose a hybrid abductive solver that autoregressively combines a dense bi-encoder with a sparse model of explanatory power, computed leveraging explicit patterns in the explanations. Our experiments demonstrate that the proposed framework can achieve performance comparable with the state-of-the-art cross-encoder while being ≈ 50 times faster and scalable to corpora of millions of facts. Moreover, we study the impact of the hybridisation on semantic drift and science question answering without additional training, showing that it boosts the quality of the explanations and contributes to improved downstream inference performance.

[1]  Peter Clark,et al.  GenericsKB: A Knowledge Base of Generic Statements , 2020, ArXiv.

[2]  Sebastian Riedel,et al.  Constructing Datasets for Multi-hop Reading Comprehension Across Documents , 2017, TACL.

[3]  Graham Neubig,et al.  Differentiable Reasoning over a Virtual Knowledge Base , 2020, ICLR.

[4]  Chitta Baral,et al.  Careful Selection of Knowledge to Solve Open Book Question Answering , 2019, ACL.

[5]  Zili Zhou,et al.  Encoding Explanatory Knowledge for Zero-shot Science Question Answering , 2021, IWCS.

[6]  Oren Etzioni,et al.  Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge , 2018, ArXiv.

[7]  Peter Clark,et al.  Can a Suit of Armor Conduct Electricity? A New Dataset for Open Book Question Answering , 2018, EMNLP.

[8]  Fabio Petroni,et al.  Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks , 2020, NeurIPS.

[9]  Wenhan Xiong,et al.  Answering Complex Open-Domain Questions with Multi-Hop Dense Retrieval , 2020, International Conference on Learning Representations.

[10]  William W. Cohen,et al.  Differentiable Open-Ended Commonsense Reasoning , 2021, NAACL.

[11]  Tushar Khot,et al.  QASC: A Dataset for Question Answering via Sentence Composition , 2020, AAAI.

[12]  Peter Jansen,et al.  What’s in an Explanation? Characterizing Knowledge and Inference Requirements for Elementary Science Exams , 2016, COLING.

[13]  Harsh Jhamtani,et al.  Learning to Explain: Datasets and Models for Identifying Valid Reasoning Chains in Multihop Question-Answering , 2020, EMNLP.

[14]  Dmitry Ustalov,et al.  TextGraphs 2019 Shared Task on Multi-Hop Inference for Explanation Regeneration , 2019, EMNLP.

[15]  Marco Valentino,et al.  ExplanationLP: Abductive Reasoning for Explainable Science Question Answering , 2020, ArXiv.

[16]  Dan Roth,et al.  On the Capabilities and Limitations of Reasoning for Natural Language Understanding , 2019, ArXiv.

[17]  Marco Valentino,et al.  Identifying Supporting Facts for Multi-hop Question Answering with Document Graph Networks , 2019, TextGraphs@EMNLP.

[18]  Mihai Surdeanu,et al.  Quick and (not so) Dirty: Unsupervised Selection of Justification Sentences for Multi-hop Question Answering , 2019, EMNLP.

[19]  Marco Valentino,et al.  Unification-based Reconstruction of Multi-hop Explanations for Science Questions , 2021, EACL.

[20]  Marco Valentino,et al.  A Survey on Explainability in Machine Reading Comprehension , 2020, ArXiv.

[21]  Jordan Boyd-Graber,et al.  Multi-Step Reasoning Over Unstructured Text with Beam Dense Retrieval , 2021, NAACL.

[22]  Yann LeCun,et al.  Dimensionality Reduction by Learning an Invariant Mapping , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[23]  Mihai Surdeanu,et al.  Unsupervised Alignment-based Iterative Evidence Retrieval for Multi-hop Question Answering , 2020, ACL.

[24]  Hugo Zaragoza,et al.  The Probabilistic Relevance Framework: BM25 and Beyond , 2009, Found. Trends Inf. Retr..

[25]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[26]  Pratyay Banerjee ASU at TextGraphs 2019 Shared Task: Explanation ReGeneration using Language Models and Iterative Re-Ranking , 2019, TextGraphs@EMNLP.

[27]  Deborah Ferreira,et al.  Natural Language Premise Selection: Finding Supporting Statements for Mathematical Text , 2020, LREC.

[28]  Danqi Chen,et al.  Dense Passage Retrieval for Open-Domain Question Answering , 2020, EMNLP.

[29]  Peter A. Jansen,et al.  WorldTree V2: A Corpus of Science-Domain Structured Explanations and Inference Patterns supporting Multi-Hop Inference , 2020, LREC.

[30]  Marie-Francine Moens,et al.  Autoregressive Reasoning over Chains of Facts with Transformers , 2020, COLING.

[31]  Clayton T. Morrison,et al.  WorldTree: A Corpus of Explanation Graphs for Elementary Science Questions supporting Multi-hop Inference , 2018, LREC.

[32]  Sam Witteveen,et al.  Red Dragon AI at TextGraphs 2019 Shared Task: Language Model Assisted Explanation Generation , 2019, TextGraphs@EMNLP.

[33]  Zhe Gan,et al.  Hierarchical Graph Network for Multi-hop Question Answering , 2020, EMNLP.

[34]  Yoshua Bengio,et al.  HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering , 2018, EMNLP.

[35]  Ming Tu,et al.  Select, Answer and Explain: Interpretable Multi-hop Reading Comprehension over Multiple Documents , 2020, AAAI.

[36]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[37]  Dan Roth,et al.  Looking Beyond the Surface: A Challenge Set for Reading Comprehension over Multiple Sentences , 2018, NAACL.

[38]  Richard Socher,et al.  Learning to Retrieve Reasoning Paths over Wikipedia Graph for Question Answering , 2019, ICLR.

[39]  Rajarshi Das,et al.  Chains-of-Reasoning at TextGraphs 2019 Shared Task: Reasoning over Chains of Facts for Explainable Multi-hop Inference , 2019, EMNLP.

[40]  Chengjie Sun,et al.  HopRetriever: Retrieve Hops over Wikipedia to Answer Complex Questions , 2020, AAAI.

[41]  Ana Marasovi'c,et al.  Teach Me to Explain: A Review of Datasets for Explainable NLP , 2021, ArXiv.

[42]  Jeff Johnson,et al.  Billion-Scale Similarity Search with GPUs , 2017, IEEE Transactions on Big Data.

[43]  Percy Liang,et al.  Transforming Question Answering Datasets Into Natural Language Inference Datasets , 2018, ArXiv.