Braid: Weaving Symbolic and Statistical Knowledge into Coherent Logical Explanations

Traditional symbolic reasoning engines, while attractive for their precision and explicability, have a few major drawbacks: the use of brittle inference procedures that rely on exact matching/unification of logical terms, an inability to deal with uncertainty, and the need for a precompiled rule-base of knowledge (the “knowledge acquisition” problem). These issues are particularly severe for the Natural Language Understanding (NLU) task, where we often use implicit background knowledge to understand and reason about text, resort to imperfect/fuzzy alignment of concepts and relations during reasoning, and constantly deal with ambiguity in representations. To address these issues, we devise a novel FOL-based reasoner, called Braid, that supports probabilistic rules, and uses the notion of custom unification functions and dynamic rule generation to overcome the brittle matching and knowledge-gap problem prevalent in traditional reasoners. In this paper, we describe the reasoning algorithms used in Braid-BC (the backchaining component of Braid), and their implementation in a distributed task-based framework that builds proof/explanation graphs for an input query in a highly scalable manner. We use a simple QA example from a children’s story to motivate Braid-BC’s design and explain how the various components work together to produce a coherent logical explanation.

[1]  Michael Kifer,et al.  HILOG: A Foundation for Higher-Order Logic Programming , 1993, J. Log. Program..

[2]  Yarden Katz,et al.  Pellet: A practical OWL-DL reasoner , 2007, J. Web Semant..

[3]  Matthew Richardson,et al.  Markov logic networks , 2006, Machine Learning.

[4]  Jennifer Chu-Carroll,et al.  To Test Machine Comprehension, Start by Defining Comprehension , 2020, ACL.

[5]  Charles L. Forgy,et al.  Rete: a fast algorithm for the many pattern/many object pattern match problem , 1991 .

[6]  Robert A. Kowalski,et al.  Linear Resolution with Selection Function , 1971, Artif. Intell..

[7]  Jennifer Chu-Carroll,et al.  GLUCOSE: GeneraLized and COntextualized Story Explanations , 2020, EMNLP.

[8]  Colin Raffel,et al.  Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..

[9]  Andrei Voronkov,et al.  First-Order Theorem Proving and Vampire , 2013, CAV.

[10]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[11]  Douglas B. Lenat,et al.  CYC: Using Common Sense Knowledge to Overcome Brittleness and Knowledge Acquisition Bottlenecks , 1986, AI Mag..

[12]  Gerard de Melo,et al.  External Sources of Axioms in Automated Theorem Proving , 2009, KI.

[13]  Wolfgang Faber Answer Set Programming , 2013, Reasoning Web.

[14]  Jennifer Chu-Carroll,et al.  Open-Domain Frame Semantic Parsing Using Transformers , 2020, ArXiv.

[15]  Josef Ruppenhofer,et al.  FrameNet II: Extended theory and practice , 2006 .

[16]  Artur S. d'Avila Garcez,et al.  Logic Tensor Networks: Deep Learning and Logical Reasoning from Data and Knowledge , 2016, NeSy@HLAI.