Braid: Weaving Symbolic and Neural Knowledge into Coherent Logical Explanations.

Traditional symbolic reasoning engines, while attractive for their precision and explicability, have a few major drawbacks: the use of brittle inference procedures that rely on exact matching (unification) of logical terms, an inability to deal with uncertainty, and the need for a precompiled rule-base of knowledge (the "knowledge acquisition" problem). These issues are particularly severe for the Natural Language Understanding (NLU) task, where we often use implicit background knowledge to understand and reason about text, resort to fuzzy alignment of concepts and relations during reasoning, and constantly deal with ambiguity in representations. To address these issues, we devise a novel FOL-based reasoner, called Braid, that supports probabilistic rules, and uses the notion of custom unification functions and dynamic rule generation to overcome the brittle matching and knowledge-gap problem prevalent in traditional reasoners. In this paper, we describe the reasoning algorithms used in Braid-BC (the backchaining component of Braid), and their implementation in a distributed task-based framework that builds proof/explanation graphs for an input query in a scalable manner. We use a simple QA example from a children's story to motivate Braid-BC's design and explain how the various components work together to produce a coherent logical explanation.

[1]  Nathanael Chambers,et al.  A Corpus and Cloze Evaluation for Deeper Understanding of Commonsense Stories , 2016, NAACL.

[2]  Robert A. Kowalski,et al.  Linear Resolution with Selection Function , 1971, Artif. Intell..

[3]  Jennifer Chu-Carroll,et al.  GLUCOSE: GeneraLized and COntextualized Story Explanations , 2020, EMNLP.

[4]  Colin Raffel,et al.  Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..

[5]  Douglas B. Lenat,et al.  CYC: Using Common Sense Knowledge to Overcome Brittleness and Knowledge Acquisition Bottlenecks , 1986, AI Mag..

[6]  Xiaoyan Zhu,et al.  Story Ending Selection by Finding Hints From Pairwise Candidate Endings , 2019, IEEE/ACM Transactions on Audio, Speech, and Language Processing.

[7]  Ulf Leser,et al.  NLProlog: Reasoning with Weak Unification for Question Answering in Natural Language , 2019, ACL.

[8]  Michael Kifer,et al.  HILOG: A Foundation for Higher-Order Logic Programming , 1993, J. Log. Program..

[9]  Matthew Richardson,et al.  Markov logic networks , 2006, Machine Learning.

[10]  Artur S. d'Avila Garcez,et al.  Logic Tensor Networks: Deep Learning and Logical Reasoning from Data and Knowledge , 2016, NeSy@HLAI.

[11]  Yarden Katz,et al.  Pellet: A practical OWL-DL reasoner , 2007, J. Web Semant..

[12]  Zhou Yu,et al.  Incorporating Structured Commonsense Knowledge in Story Completion , 2018, AAAI.

[13]  Marvin Minsky,et al.  A framework for representing knowledge , 1974 .

[14]  Gerard de Melo,et al.  External Sources of Axioms in Automated Theorem Proving , 2009, KI.

[15]  Wolfgang Faber Answer Set Programming , 2013, Reasoning Web.

[16]  Charles L. Forgy,et al.  Rete: a fast algorithm for the many pattern/many object pattern match problem , 1991 .

[17]  Josef Ruppenhofer,et al.  FrameNet II: Extended theory and practice , 2006 .

[18]  Charles L. Forgy,et al.  Rete: A Fast Algorithm for the Many Patterns/Many Objects Match Problem , 1982, Artif. Intell..

[19]  Ting Liu,et al.  Story Ending Prediction by Transferable BERT , 2019, IJCAI.

[20]  Jennifer Chu-Carroll,et al.  Open-Domain Frame Semantic Parsing Using Transformers , 2020, ArXiv.

[21]  Jennifer Chu-Carroll,et al.  To Test Machine Comprehension, Start by Defining Comprehension , 2020, ACL.

[22]  Andrei Voronkov,et al.  First-Order Theorem Proving and Vampire , 2013, CAV.

[23]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[24]  Ilya Sutskever,et al.  Language Models are Unsupervised Multitask Learners , 2019 .