Compositionality in rational analysis: grammar-based induction for concept learning

Rational analysis attempts to explain aspects of human cognition as an adaptive response to the environment (Marr, 1982; Anderson, 1990; Chater, Tenenbaum, & Yuille, 2006). The dominant approach to rational analysis today takes an ecologically reasonable specification of a problem facing an organism, given in statistical terms, then seeks an optimal solution, usually using Bayesian methods. This approach has proven very successful in cognitive science; it has predicted perceptual phenomena (Geisler & Kersten, 2002; Feldman, 2001), illuminated puzzling effects in reasoning (Chater & Oaksford, 1999; Griffiths & Tenenbaum, 2006), and, especially, explained how human learning can succeed despite sparse input and endemic uncertainty (Tenenbaum, 1999; Tenenbaum & Griffiths, 2001). However, there were earlier notions of the “rational” analysis of cognition that emphasized very different ideas. One of the central ideas behind logical and computational approaches, which previously dominated notions of rationality, is that meaning can be captured in the structure of representations, but that compositional semantics are needed for these representations to provide a coherent account of thought. In this chapter we attempt to reconcile the modern approach to rational analysis with some aspects of this older, logico-computational approach. We do this via a model—offered as an extended example—of human concept learning. In the current chapter we are primarily concerned with formal aspects of this approach; in other work (Goodman, Tenenbaum, Feldman, & Griffiths, in press) we more carefully study a variant of this model as a psychological model of human concept learning. Explaining human cognition was one of the original motivations for the development of formal logic. George Boole, the father of digital logic, developed his symbolic language in order to explicate the rational laws underlying thought: his principal work, An Investigation of the Laws of Thought (Boole, 1854), was written to “investigate the fundamental laws of those operations of the mind by which reasoning is performed,” and arrived at “some probable intimations concerning the nature and constitution of the human mind” (p. 1). Much of mathematical logic since Boole can be regarded as an attempt to capture the coherence of thought in a formal system. This is particularly apparent in the work, by Frege (1892), Tarski (1956) and others, on model-theoretic semantics for logic, which aimed to create formal systems both flexible and systematic enough to capture the complexities of mathematical thought. A central component in this program is compositionality. Consider Frege’s Principle1: each syntactic operation of a formal language should have a corresponding semantic operation. This principle requires syntactic compositionality, that meaningful terms in a formal system are built up by combination operations, as well as compatibility between the syntax and semantics of the system. When Turing, Church, and others suggested that formal systems could be manipulated by mechanical computers it was natural (at least in hindsight) to suggest that cognition operates in a similar way: meaning is manipulated in the mind by computation2. Viewing the mind as a formal computational system in this way suggests that compositionality should also be found in the mind; that is, that mental representations may be combined into new representations, and the meaning of mental representations may be decomposed in terms of the meaning of their components. Two important virtues for a theory of thought result (Fodor, 1975): productivity—the number of representations is unbounded because they may be boundlessly combined— and systematicity—the combination of two representations is meaningful to one who can understand each separately. Despite its importance to the computational theory of mind, compositionality has seldom been captured by modern rational analyses. Yet there are a number of reasons to desire a compositional rational analysis. For instance, productivity of mental representations would provide an explanation of the otherwise puzzling ability of human thought to adapt to novel situations populated by new concepts—even those far beyond the ecological pressures of our evolutionary milieu (such as radiator repairs and the use of fiberglass bottom powerboats). We will show in this chapter that Bayesian statistical methods can be fruitfully combined with compositional representational systems by developing such a model in the well-studied setting of concept learning. This addresses a long running tension in the literature on human concepts: similarity-based statistical learning models have provided a good understanding of how simple concepts can be learned (Medin & Schaffer, 1978; Anderson, 1991; Kruschke, 1992;

[1]  J. Kruschke,et al.  ALCOVE: an exemplar-based connectionist model of category learning. , 1992, Psychological review.

[2]  N. Chater,et al.  Ten years of the rational analysis of cognition , 1999, Trends in Cognitive Sciences.

[3]  Thomas L. Griffiths,et al.  Learning Systems of Concepts with an Infinite Relational Model , 2006, AAAI.

[4]  J. Tenenbaum,et al.  Optimal Predictions in Everyday Cognition , 2006, Psychological science.

[5]  J. Tenenbaum,et al.  Generalization, similarity, and Bayesian inference. , 2001, The Behavioral and brain sciences.

[6]  David M. Sobel,et al.  A theory of causal learning in children: causal maps and Bayes nets. , 2004, Psychological review.

[7]  P. Johnson-Laird,et al.  Mental Models: Towards a Cognitive Science of Language, Inference, and Consciousness , 1985 .

[8]  J. Tenenbaum,et al.  Word learning as Bayesian inference. , 2007, Psychological review.

[9]  David M. Sobel,et al.  Detecting blickets: how young children use information about novel causal powers in categorization and induction. , 2000, Child development.

[10]  G. Lakoff,et al.  Women, Fire, and Dangerous Things: What Categories Reveal about the Mind , 1988 .

[11]  Herbert B. Enderton,et al.  A mathematical introduction to logic , 1972 .

[12]  Robert L. Goldstone,et al.  Relational similarity and the nonindependence of features in similarity judgments , 1991, Cognitive Psychology.

[13]  F. Keil,et al.  A Characteristic-to-Defining Shift in the Development of Word Meaning. , 1984 .

[14]  Christopher D. Manning,et al.  Probabilistic models of language processing and acquisition , 2006, Trends in Cognitive Sciences.

[15]  D. K. Nelson,et al.  Principle-Based Inferences in Young Children's Categorization: Revisiting the Impact of Function on the Naming of Artifacts. , 1995 .

[16]  G. Boole An Investigation of the Laws of Thought: On which are founded the mathematical theories of logic and probabilities , 2007 .

[17]  Joshua B. Tenenbaum,et al.  Inferring causal networks from observations and interventions , 2003, Cogn. Sci..

[18]  Stuart J. Russell,et al.  General-Purpose MCMC Inference over Relational Structures , 2006, UAI.

[19]  John R. Anderson,et al.  The Adaptive Nature of Human Categorization. , 1991 .

[20]  Robert L. Goldstone,et al.  Conceptual interrelatedness and caricatures , 2003, Memory & cognition.

[21]  Arthur B. Markman,et al.  Role-governed categories , 2001, J. Exp. Theor. Artif. Intell..

[22]  A. Yuille,et al.  Opinion TRENDS in Cognitive Sciences Vol.10 No.7 July 2006 Special Issue: Probabilistic models of cognition Vision as Bayesian inference: analysis by synthesis? , 2022 .

[23]  Refractor Vision , 2000, The Lancet.

[24]  Joseph Y. Halpern,et al.  Causes and explanations: A structural-model approach , 2000 .

[25]  Keith A. Markus,et al.  Making Things Happen: A Theory of Causal Explanation , 2007 .

[26]  S. Sloman,et al.  Learning Causal Structure , 2019, Proceedings of the Twenty-Fourth Annual Conference of the Cognitive Science Society.

[27]  J. Tenenbaum,et al.  Opinion TRENDS in Cognitive Sciences Vol.10 No.7 July 2006 Special Issue: Probabilistic models of cognition Theory-based Bayesian models of inductive learning and reasoning , 2022 .

[28]  Jennifer A. Asmuth,et al.  Context Sensitivity of Relational Nouns , 2005 .

[29]  Douglas L. Medin,et al.  Context theory of classification learning. , 1978 .

[30]  Steven A. Sloman,et al.  Feature Centrality and Conceptual Coherence , 1998, Cogn. Sci..

[31]  R. Nosofsky,et al.  Rule-plus-exception model of classification learning. , 1994, Psychological review.

[32]  A. Gopnik,et al.  Causal learning : psychology, philosophy, and computation , 2007 .

[33]  R. Baierlein Probability Theory: The Logic of Science , 2004 .

[34]  John R. Anderson,et al.  The Adaptive Character of Thought , 1990 .

[35]  Edward E. Smith,et al.  On the adequacy of prototype theory as a theory of concepts , 1981, Cognition.

[36]  Bob Rehder,et al.  A Causal-Model Theory of Categorization , 1999 .

[37]  B. Love Comparing supervised and unsupervised category learning , 2002, Psychonomic bulletin & review.

[38]  A. Markman,et al.  Categorization inside and outside the laboratory : essays in honor of Douglas L. Medin , 2005 .

[39]  Ludwig Wittgenstein,et al.  Tractatus Logico Philosophicus (Routledge Classics) (Routledge Classics) , 2001 .

[40]  J. Tenenbaum,et al.  Intuitive theories as grammars for causal inference , 2007 .

[41]  D. Medin,et al.  The role of theories in conceptual coherence. , 1985, Psychological review.

[42]  D. Medin,et al.  SUSTAIN: a network model of category learning. , 2004, Psychological review.

[43]  Jacob feldman,et al.  Bayesian contour integration , 2001, Perception & psychophysics.

[44]  Thomas L. Griffiths,et al.  A Rational Analysis of Rule-Based Concept Learning , 2008, Cogn. Sci..

[45]  Mark W. Altom,et al.  Correlated symptoms and simulated medical classification. , 1982, Journal of experimental psychology. Learning, memory, and cognition.

[46]  David B. Dunson,et al.  Bayesian Data Analysis , 2010 .

[47]  Stephen Muggleton,et al.  Learning from Positive Data , 1996, Inductive Logic Programming Workshop.

[48]  Richard Montague,et al.  The Proper Treatment of Quantification in Ordinary English , 1973 .

[49]  Jacob Feldman,et al.  Minimization of Boolean complexity in human concept learning , 2000, Nature.

[50]  J. Pearl Causality: Models, Reasoning and Inference , 2000 .

[51]  B. Love,et al.  Beyond common features: The role of roles in determining similarity , 2007, Cognitive Psychology.

[52]  D. Kersten,et al.  Illusions, perception and Bayes , 2002, Nature Neuroscience.

[53]  M. Posner,et al.  On the genesis of abstract ideas. , 1968, Journal of experimental psychology.

[54]  W. Ahn,et al.  Causal Status as a Determinant of Feature Centrality , 2000, Cognitive Psychology.

[55]  J. Feldman An algebra of human concept learning , 2006 .

[56]  J. Tenenbaum,et al.  Special issue on “Probabilistic models of cognition , 2022 .

[57]  J. Tenenbaum A Bayesian framework for concept learning , 1999 .

[58]  G. Frege Über Sinn und Bedeutung , 1892 .

[59]  E. Rosch,et al.  Categorization of Natural Objects , 1981 .