Navigating the Semantic Horizon using Relative Neighborhood Graphs

This paper is concerned with nearest neighbor search in distributional semantic models. A normal nearest neighbor search only returns a ranked list of neighbors, with no information about the structure or topology of the local neighborhood. This is a potentially serious shortcoming of the mode of querying a distributional semantic model, since a ranked list of neighbors may conflate several different senses. We argue that the topology of neighborhoods in semantic space provides important information about the different senses of terms, and that such topological structures can be used for word-sense induction. We also argue that the topology of the neighborhoods in semantic space can be used to determine the semantic horizon of a point, which we define as the set of neighbors that have a direct connection to the point. We introduce relative neighborhood graphs as method to uncover the topological properties of neighborhoods in semantic models. We also provide examples of relative neighborhood graphs for three well-known semantic models; the PMI model, the GloVe model, and the skipgram model.

[1]  Hitoshi Isahara,et al.  Clustering Using Feature Domain Similarity to Discover Word Senses for Adjectives , 2007, International Conference on Semantic Computing (ICSC 2007).

[2]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[3]  Katrin Erk,et al.  Exemplar-Based Models for Word Meaning in Context , 2010, ACL.

[4]  Raymond J. Mooney,et al.  Multi-Prototype Vector-Space Models of Word Meaning , 2010, NAACL.

[5]  Jean Véronis,et al.  HyperLex: lexical cartography for information retrieval , 2004, Comput. Speech Lang..

[6]  Hamish Carr,et al.  Topological Methods in Data Analysis and Visualization III, Theory, Algorithms, and Applications , 2011 .

[7]  Peter Lindstrom,et al.  Locally-scaled spectral clustering using empty region graphs , 2012, KDD.

[8]  Mirella Lapata,et al.  Measuring Distributional Similarity in Context , 2010, EMNLP.

[9]  Suresh Manandhar,et al.  Word Sense Induction Using Graphs of Collocations , 2008, ECAI.

[10]  Ted Pedersen Duluth-WSI: SenseClusters Applied to the Sense Induction Task of SemEval-2 , 2010, SemEval@ACL.

[11]  Ted Pedersen,et al.  Word Sense Discrimination by Clustering Contexts in Vector and Similarity Spaces , 2004, CoNLL.

[12]  Eneko Agirre,et al.  Semeval-2007 Task 2 : Evaluating Word Sense Induction and Discrimination , 2007 .

[13]  Timothy Baldwin,et al.  Word Sense Induction for Novel Sense Detection , 2012, EACL.

[14]  Piotr Indyk,et al.  Approximate Nearest Neighbor: Towards Removing the Curse of Dimensionality , 2012, Theory Comput..

[15]  Xuchen Yao,et al.  Nonparametric Bayesian Word Sense Induction , 2011, Graph-based Methods for Natural Language Processing.

[16]  S. M. García,et al.  2014: , 2020, A Party for Lazarus.

[17]  Magnus Sahlgren,et al.  Filaments of Meaning in Word Space , 2008, ECIR.

[18]  Jean Cardinal,et al.  Empty region graphs , 2009, Comput. Geom..

[19]  Roberto Navigli,et al.  Clustering and Diversifying Web Search Results with Graph-Based Word Sense Induction , 2013, CL.

[20]  Anna Korhonen,et al.  Probabilistic models of similarity in syntactic context , 2011, EMNLP.

[21]  Naftali Tishby,et al.  Distributional Clustering of English Words , 1993, ACL.

[22]  Godfried T. Toussaint,et al.  The relative neighbourhood graph of a finite planar set , 1980, Pattern Recognit..

[23]  Patrick Pantel,et al.  Discovering word senses from text , 2002, KDD.

[24]  Keith Stevens,et al.  HERMIT: Flexible Clustering for the SemEval-2 WSI Task , 2010, SemEval@ACL.

[25]  Mirella Lapata,et al.  Bayesian Word Sense Induction , 2009, EACL.

[26]  Eneko Agirre,et al.  Two graph-based algorithms for state-of-the-art WSD , 2006, EMNLP.

[27]  DAYID,et al.  David Newman , 1924, Glasgow Medical Journal.

[28]  Omer Levy,et al.  Neural Word Embedding as Implicit Matrix Factorization , 2014, NIPS.

[29]  Sunil Arya,et al.  An optimal algorithm for approximate nearest neighbor searching fixed dimensions , 1998, JACM.

[30]  Sergey Brin,et al.  The Anatomy of a Large-Scale Hypertextual Web Search Engine , 1998, Comput. Networks.

[31]  S. Dongen Graph clustering by flow simulation , 2000 .

[32]  Marianna Apidianaki,et al.  Latent Semantic Word Sense Induction and Disambiguation , 2011, ACL.

[33]  Erik Velldal,et al.  A Fuzzy Clustering Approach to Word Sense Discrimination , 2005 .

[34]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[35]  Christian Biemann,et al.  Chinese Whispers - an Efficient Graph Clustering Algorithm and its Application to Natural Language Processing Problems , 2006 .

[36]  Roberto Navigli,et al.  SemEval-2013 Task 11: Word Sense Induction and Disambiguation within an End-User Application , 2013, SemEval@NAACL-HLT.

[37]  Suresh Manandhar,et al.  SemEval-2010 Task 14: Word Sense Induction &Disambiguation , 2010, SemEval@ACL.

[38]  David Jurgens,et al.  SemEval-2013 Task 13: Word Sense Induction for Graded and Non-Graded Senses , 2013, SemEval@NAACL-HLT.

[39]  Hinrich Schütze,et al.  Automatic Word Sense Discrimination , 1998, Comput. Linguistics.

[40]  Piotr Indyk,et al.  Approximate nearest neighbors: towards removing the curse of dimensionality , 1998, STOC '98.

[41]  Roberto Navigli,et al.  Paving the Way to a Large-scale Pseudosense-annotated Dataset , 2013, HLT-NAACL.

[42]  Jon Louis Bentley,et al.  Multidimensional binary search trees used for associative searching , 1975, CACM.

[43]  BentleyJon Louis Multidimensional binary search trees used for associative searching , 1975 .

[44]  Dominic Widdows,et al.  Discovering Corpus-Specific Word Senses , 2003, EACL.