Modelling Semantic Categories using Conceptual Neighborhood

While many methods for learning vector space embeddings have been proposed in the field of Natural Language Processing, these methods typically do not distinguish between categories and individuals. Intuitively, if individuals are represented as vectors, we can think of categories as (soft) regions in the embedding space. Unfortunately, meaningful regions can be difficult to estimate, especially since we often have few examples of individuals that belong to a given category. To address this issue, we rely on the fact that different categories are often highly interdependent. In particular, categories often have conceptual neighbors, which are disjoint from but closely related to the given category (e.g. fruit and vegetable). Our hypothesis is that more accurate category representations can be learned by relying on the assumption that the regions representing such conceptual neighbors should be adjacent in the embedding space. We propose a simple method for identifying conceptual neighbors and then show that incorporating these conceptual neighbors indeed leads to more accurate region based representations.

[1]  Travé-Massuyès Conceptual Neighborhood and its role in temporal and spatial reasoning , 1991 .

[2]  Gemma Boleda,et al.  Instances and concepts in distributional space , 2017, EACL.

[3]  Erik Cambria,et al.  Label Embedding for Zero-shot Fine-grained Named Entity Typing , 2016, COLING.

[4]  Steven Schockaert,et al.  Entity Embeddings with Conceptual Subspaces as a Basis for Plausible Reasoning , 2016, ECAI.

[5]  Robert L. Goldstone Isolated and interrelated concepts , 1996, Memory & cognition.

[6]  James Pustejovsky,et al.  Towards a Generative Lexical Resource: The Brandeis Semantic Ontology , 2006, LREC.

[7]  Rajarshi Das,et al.  Gaussian LDA for Topic Models with Word Embeddings , 2015, ACL.

[8]  Tiziano Flati,et al.  MultiWiBi: The multilingual Wikipedia bitaxonomy project , 2016, Artif. Intell..

[9]  Ellen M. Markman,et al.  Constraints Children Place on Word Meanings , 1990, Cogn. Sci..

[10]  Steven Schockaert,et al.  MEmbER: Max-Margin Based Embeddings for Entity Retrieval , 2017, SIGIR.

[11]  Steven Schockaert,et al.  Modeling Context Words as Regions: An Ordinal Regression Approach to Word Embedding , 2017, CoNLL.

[12]  Gemma Boleda,et al.  Inclusive yet Selective: Supervised Distributional Hypernymy Detection , 2014, COLING.

[13]  Zhiting Hu,et al.  Joint Embedding of Hierarchical Categories and Entities for Concept Categorization and Dataless Classification , 2016, COLING.

[14]  Steven Schockaert,et al.  Inductive Reasoning about Ontologies Using Conceptual Spaces , 2017, AAAI.

[15]  Richard A. Harshman,et al.  Indexing by Latent Semantic Analysis , 1990, J. Am. Soc. Inf. Sci..

[16]  Dipanjan Das,et al.  BERT Rediscovers the Classical NLP Pipeline , 2019, ACL.

[17]  James Pustejovsky,et al.  Classification of Discourse Coherence Relations: An Exploratory Study using Multiple Knowledge Sources , 2006, SIGDIAL Workshop.

[18]  Steven Schockaert,et al.  Learning Conceptual Space Representations of Interrelated Concepts , 2018, IJCAI.

[19]  Jason Weston,et al.  Translating Embeddings for Modeling Multi-relational Data , 2013, NIPS.

[20]  Ming-Wei Chang,et al.  Inferring Missing Entity Type Instances for Knowledge Base Completion: New Dataset and Methods , 2015, NAACL.

[21]  Andrew McCallum,et al.  Word Representations via Gaussian Embedding , 2014, ICLR.

[22]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[23]  David J. Weir,et al.  Learning to Distinguish Hypernyms and Co-Hyponyms , 2014, COLING.

[24]  Simone Paolo Ponzetto,et al.  BabelNet: The automatic construction, evaluation and application of a wide-coverage multilingual semantic network , 2012, Artif. Intell..

[25]  N. Foo Conceptual Spaces—The Geometry of Thought , 2022 .

[26]  Katrin Erk,et al.  Representing words as regions in vector space , 2009, CoNLL.

[27]  Ignacio Iacobacci,et al.  Embedding Words and Senses Together via Joint Knowledge-Enhanced Training , 2016, CoNLL.

[28]  Sanjeev Arora,et al.  A Simple but Tough-to-Beat Baseline for Sentence Embeddings , 2017, ICLR.

[29]  Roberto Navigli,et al.  Nasari: Integrating explicit knowledge and corpus statistics for a multilingual representation of concepts and entities , 2016, Artif. Intell..

[30]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[31]  Luke S. Zettlemoyer,et al.  Deep Contextualized Word Representations , 2018, NAACL.

[32]  Luke S. Zettlemoyer,et al.  Dissecting Contextual Word Embeddings: Architecture and Representation , 2018, EMNLP.

[33]  Markus Krötzsch,et al.  Wikidata , 2014, Commun. ACM.

[34]  Nematollah Batmanghelich,et al.  Nonparametric Spherical Topic Modeling with Word Embeddings , 2016, ACL.

[35]  Li Guo,et al.  Semantically Smooth Knowledge Graph Embedding , 2015, ACL.

[36]  Steven Schockaert,et al.  Reasoning about Betweenness and RCC8 Constraints in Qualitative Conceptual Spaces , 2018, IJCAI.

[37]  James Pustejovsky,et al.  The Generative Lexicon , 1995, CL.

[38]  Gemma Boleda,et al.  Distributional vectors encode referential attributes , 2015, EMNLP.