Towards General Natural Language Understanding with Probabilistic Worldbuilding

We introduce the Probabilistic Worldbuilding Model (PWM), a new fully-symbolic Bayesian model of semantic parsing and reasoning, as a first step in a research program toward more domainand task-general NLU and AI. Humans create internal mental models of their observations which greatly aid in their ability to understand and reason about a large variety of problems. In PWM, the meanings of sentences, acquired facts about the world, and intermediate steps in reasoning are all expressed in a humanreadable formal language, with the design goal of interpretability. PWM is Bayesian, designed specifically to be able to generalize to new domains and new tasks. We derive and implement an inference algorithm that reads sentences by parsing and abducing updates to its latent world model that capture the semantics of those sentences, and evaluate it on two out-of-domain questionanswering datasets: (1) ProofWriter and (2) a new dataset we call FictionalGeoQA, designed to be more representative of real language but still simple enough to focus on evaluating reasoning ability, while being robust against heuristics. Our method outperforms baselines on both, thereby demonstrating its value as a proof-of-concept.

[1]  Robert P. Goldman,et al.  A Bayesian Model of Plan Recognition , 1993, Artif. Intell..

[2]  John K. Tsotsos,et al.  40 years of cognitive architectures: core cognitive abilities and practical applications , 2018, Artificial Intelligence Review.

[3]  Luc De Raedt,et al.  Scalable Rule Learning in Probabilistic Knowledge Bases , 2019, AKBC.

[4]  Rolf Morel,et al.  Learning programs by learning from failures , 2020, Mach. Learn..

[5]  James Cussens,et al.  Parameter Estimation in Stochastic Logic Programs , 2001, Machine Learning.

[6]  Terence Parsons,et al.  Events in the Semantics of English: A Study in Subatomic Semantics , 1990 .

[7]  Tom Mitchell,et al.  A Probabilistic Generative Grammar for Semantic Parsing , 2016, CoNLL.

[8]  Andrei Voronkov,et al.  First-Order Theorem Proving and Vampire , 2013, CAV.

[9]  Neng-Fa Zhou,et al.  Generative Modeling with Failure in PRISM , 2005, IJCAI.

[10]  Mohit Bansal,et al.  PRover: Proof Generation for Interpretable Reasoning over Rules , 2020, EMNLP.

[11]  M. E. Szabo,et al.  The collected papers of Gerhard Gentzen , 1969 .

[12]  Raymond J. Mooney,et al.  Learning to Parse Database Queries Using Inductive Logic Programming , 1996, AAAI/IAAI, Vol. 2.

[13]  Miriam R. L. Petruck,et al.  Language (Re)modelling: Towards Embodied Language Understanding , 2020, ACL.

[14]  Tal Linzen,et al.  How Can We Accelerate Progress Towards Human-like Linguistic Generalization? , 2020, ACL.

[15]  Heiner Stuckenschmidt,et al.  Towards Distributed MCMC Inference in Probabilistic Knowledge Bases , 2012, AKBC-WEKEX@NAACL-HLT.

[16]  Xinlei Chen,et al.  Never-Ending Learning , 2012, ECAI.

[17]  Yiming Yang,et al.  XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.

[18]  T. Ferguson A Bayesian Analysis of Some Nonparametric Problems , 1973 .

[19]  W. V. Quine,et al.  Natural deduction , 2021, An Introduction to Proof Theory.

[20]  Jure Leskovec,et al.  Query2box: Reasoning over Knowledge Graphs in Vector Space using Box Embeddings , 2020, ICLR.

[21]  Simon Cruanes,et al.  Faster, Higher, Stronger: E 2.3 , 2019, CADE.

[22]  Staffan Larsson,et al.  Probabilistic Type Theory and Natural Language Semantics , 2015, LILT.

[23]  Allen Newell,et al.  Computer science as empirical inquiry: symbols and search , 1976, CACM.

[24]  Mark Chen,et al.  Language Models are Few-Shot Learners , 2020, NeurIPS.

[25]  Emily M. Bender,et al.  Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data , 2020, ACL.

[26]  William W. Cohen,et al.  Faithful Embeddings for Knowledge Base Queries , 2020, NeurIPS.

[27]  Thomas L. Griffiths,et al.  Adaptor Grammars: A Framework for Specifying Compositional Nonparametric Bayesian Models , 2006, NIPS.

[28]  Vítor Santos Costa,et al.  Inductive Logic Programming , 2013, Lecture Notes in Computer Science.

[29]  Christian P. Robert,et al.  Monte Carlo Statistical Methods (Springer Texts in Statistics) , 2005 .

[30]  Ulrich Furbach,et al.  Tackling Benchmark Problems of Commonsense Reasoning , 2015, Bridging@CADE.

[31]  Fabrizio Riguzzi,et al.  Structure learning of probabilistic logic programs by searching the clause space , 2013, Theory and Practice of Logic Programming.

[32]  Jonathan Berant,et al.  On Making Reading Comprehension More Comprehensive , 2019, EMNLP.

[33]  Chuang Gan,et al.  CLEVRER: CoLlision Events for Video REpresentation and Reasoning , 2020, ICLR.

[34]  Colin Raffel,et al.  Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..

[35]  Ryan P. Adams,et al.  SUMO: Unbiased Estimation of Log Marginal Probability for Latent Variable Models , 2020, ICLR.

[36]  G. Gentzen Untersuchungen über das logische Schließen. I , 1935 .

[37]  Jennifer Chu-Carroll,et al.  To Test Machine Comprehension, Start by Defining Comprehension , 2020, ACL.

[38]  Peter Clark,et al.  ProofWriter: Generating Implications, Proofs, and Abductive Statements over Natural Language , 2020, FINDINGS.

[39]  Leon Henkin,et al.  Completeness in the theory of types , 1950, Journal of Symbolic Logic.

[40]  Johan Bos,et al.  Open-Domain Semantic Parsing with Boxer , 2015, NODALIDA.

[41]  S. Muggleton Stochastic Logic Programs , 1996 .

[42]  Ailsa H. Land,et al.  An Automatic Method of Solving Discrete Programming Problems , 1960 .

[43]  Joshua B. Tenenbaum,et al.  Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.

[44]  Jerry R. Hobbs,et al.  Interpretation as Abduction , 1993, Artif. Intell..

[45]  Howard Gregory Language and Logics: An Introduction to the Logical Foundations of Language , 2015 .

[46]  Allen Newell,et al.  SOAR: An Architecture for General Intelligence , 1987, Artif. Intell..

[47]  David R. Dowty,et al.  Introduction to Montague semantics , 1980 .

[48]  Hoon Kim,et al.  Monte Carlo Statistical Methods , 2000, Technometrics.

[49]  Pedro M. Domingos,et al.  Learning and Inference in Tractable Probabilistic Knowledge Bases , 2015, UAI.

[50]  D. Aldous Exchangeability and related topics , 1985 .

[51]  Gerhard Gentzen,et al.  Investigations into Logical Deduction , 1970 .

[52]  Alonzo Church,et al.  A formulation of the simple theory of types , 1940, Journal of Symbolic Logic.

[53]  Hannaneh Hajishirzi,et al.  UnifiedQA: Crossing Format Boundaries With a Single QA System , 2020, FINDINGS.

[54]  Doug Downey,et al.  Abductive Commonsense Reasoning , 2019, ICLR.

[55]  Stuart J. Russell,et al.  Artificial Intelligence - A Modern Approach, Third International Edition , 2010 .

[56]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[57]  Steffen Staab,et al.  Knowledge graphs , 2021, Commun. ACM.

[58]  Pasquale Minervini,et al.  Complex Query Answering with Neural Link Predictors , 2021, ICLR.

[59]  Johan Bos,et al.  The Parallel Meaning Bank: Towards a Multilingual Corpus of Translations Annotated with Compositional Meaning Representations , 2017, EACL.

[60]  Omer Levy,et al.  RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.

[61]  W. K. Hastings,et al.  Monte Carlo Sampling Methods Using Markov Chains and Their Applications , 1970 .

[62]  Tim Rocktäschel,et al.  End-to-end Differentiable Proving , 2017, NIPS.

[63]  Dan Klein,et al.  Type-Based MCMC , 2010, HLT-NAACL.

[64]  Peter Clark,et al.  Transformers as Soft Reasoners over Language , 2020, ArXiv.