Geometric Ordering of Concepts, Logical Disjunction, and Learning by Induction

In many of the abstract geometric models which have been used to represent concepts and their relationships, regions possessing some cohesive property such as convexity or linearity have played a significant role. When the implication or containment relationship is used as an ordering relationship in such models, this gives rise to logical operators for which the disjunction of two concepts is often larger than the set union obtained in Boolean models. This paper describes some of the characteristic properties of such broad non-distributive composition operations and their applications to learning algorithms and classification structures. As an example we describe a quad-tree representation which we have used to provide a structure for indexing objects and composition of regions in a spatial database. The quad-tree combines logical, algebraic and geometric properties in a naturally non-distributive fashion. The lattice of subspaces of a vector space is presented as a special example, which draws a middle-way between ‘noninductive’ Boolean logic and ‘overinductive’ tree-structures. This gives rise to composition operations that are already used as models in physics and cognitive science. Closure conditions in geometric models The hypothesis that concepts can be represented by points and more general regions in spatial models has been used by psychologists and cognitive scientists to simulate human language learning (Landauer & Dumais 1997) and to represent sensory stimuli such as tastes and colors (Gardenfors 2000, §1.5). Of the traditional practical applications of such a spatial approach, the vector space model for information retrieval (Salton & McGill 1983) is notable, and its generalizations such as latent semantic analysis (Landauer & Dumais 1997), in which distributions of word usage learned from corpora become condensed into a lowdimensional representation and used, among other things, for discriminating between different senses of ambiguous words (Schutze 1998). Schutze’s (1998) paper exemplifies some of the opportunities and challenges involved in such a spatial approach — these include learning to represent individual objects as Copyright c © 2004, American Association for Artificial Intelligence (www.aaai.org). All rights reserved. Figure 1: Two non-convex sets (dark gray) and the points added to form their convex closures (light gray) points in a geometric space (in this case, word vectors), combining these points into appropriate sentence or document vectors (in this case, using addition of vectors), and extrapolating from observed points of information to apportion the geometric space into cohesive regions corresponding to recognizable concepts (in this case, using clustering). The last question — how are empirical observations gathered into classes described by the same word or represented by the same concept? — is of traditional importance in philosophy and many related disciplines. The extrapolation from observed data to classifying previously unexperienced situations is implemented in a variety of theoretical models and practical applications, using smoothing and clustering, by exploiting a natural general-to-specific ordering on the space of observations (Mitchell 1997, Ch. 6, 7, 2), and by using similarity or distance measures to gauge the influence exerted by a cluster of observations upon its conceptual hinterland (Gardenfors 2000, Ch. 3,4). Mathematically, such extrapolation techniques are related to closure conditions, a set being closed if it has no tendency to include new members. A traditional example of closure is in the field of topology, which describes a set as being closed if it contains the limit point of every possible sequence of elements. A more easily-grasped example is given by the property of convexity. A set S is said to be convex if for any two pointsA andB in S, the straight lineAB lies entirely within S. The convex closure of S is formed by taking the initial set and all such straight lines, this being the smallest convex set containing S. Figure 1 shows two simple non-convex regions and their convex closures. One of the best developed uses of such closure methods for obtaining stable conceptual representations is in Formal Figure 2: The convex closure of the union of two sets. Concept Analysis, where the closure operation is given by the relationship between the intent and the extent of a concept (Ganter & Wille 1999, §1.1). An important closure operation we will consider later is the linear span of a set of vectors, which can also be thought of as the smallest subspace containing those vectors. Ordering, containment, implication and

[1]  Joan Lasenby,et al.  A unified mathematical language for physics and engineering in the 21st century , 2000, Philosophical Transactions of the Royal Society of London. Series A: Mathematical, Physical and Engineering Sciences.

[2]  Bernhard Ganter,et al.  Formal Concept Analysis: Mathematical Foundations , 1998 .

[3]  Hinrich Schütze,et al.  Automatic Word Sense Discrimination , 1998, Comput. Linguistics.

[4]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[5]  Hinrich Sch Automatic Word Sense Discrimination , 1998 .

[6]  Michael McGill,et al.  Introduction to Modern Information Retrieval , 1983 .

[7]  Hanan Samet,et al.  Spatial Databases , 1992, VLDB.

[8]  John E. Howland,et al.  Computer graphics , 1990, IEEE Potentials.

[9]  N. Foo Conceptual Spaces—The Geometry of Thought , 2022 .

[10]  Hinrich Schütze,et al.  Book Reviews: Foundations of Statistical Natural Language Processing , 1999, CL.

[11]  T. Landauer,et al.  A Solution to Plato's Problem: The Latent Semantic Analysis Theory of Acquisition, Induction, and Representation of Knowledge. , 1997 .

[12]  Dominic Widdows,et al.  Geometry and Meaning , 2004, Computational Linguistics.

[13]  Dominic Widdows,et al.  Orthogonal Negation in Vector Spaces for Modelling Word-Meanings and Document Retrieval , 2003, ACL.

[14]  Yannis Manolopoulos,et al.  Spatial Databases , 2004 .

[15]  Tony A. Plate,et al.  Holographic Reduced Representation: Distributed Representation for Cognitive Structures , 2003 .

[16]  Hans-Peter Kriegel,et al.  The R*-tree: an efficient and robust access method for points and rectangles , 1990, SIGMOD '90.

[17]  H. Dishkant,et al.  Logic of Quantum Mechanics , 1976 .

[18]  George Boole,et al.  An Investigation of the Laws of Thought: Frontmatter , 2009 .

[19]  R. S. Ward,et al.  Twistor Geometry and Field Theory , 1990 .

[20]  Hang Li,et al.  Generalizing Case Frames Using a Thesaurus and the MDL Principle , 1995, CL.

[21]  George Boole,et al.  The mathematical analysis of logic , 1948 .