Evaluating Neighbor Rank and Distance Measures as Predictors of Semantic Priming

This paper summarizes the results of a large-scale evaluation study of bag-ofwords distributional models on behavioral data from three semantic priming experiments. The tasks at issue are (i) identification of consistent primes based on their semantic relatedness to the target and (ii) correlation of semantic relatedness with latency times. We also provide an evaluation of the impact of specific model parameters on the prediction of priming. To the best of our knowledge, this is the first systematic evaluation of a wide range of DSM parameters in all possible combinations. An important result of the study is that neighbor rank performs better than distance measures in predicting semantic priming.

[1]  W. Lowe,et al.  The Direct Route: Mediated Priming in Semantic Space , 2000 .

[2]  T. Landauer,et al.  A Solution to Plato's Problem: The Latent Semantic Analysis Theory of Acquisition, Induction, and Representation of Knowledge. , 1997 .

[3]  Magnus Sahlgren,et al.  An Introduction to Random Indexing , 2005 .

[4]  A. Tversky Features of Similarity , 1977 .

[5]  Stefan Evert,et al.  The Statistics of Word Cooccur-rences: Word Pairs and Collocations , 2004 .

[6]  Mirella Lapata,et al.  Vector-based Models of Semantic Composition , 2008, ACL.

[7]  Patrick Pantel,et al.  From Frequency to Meaning: Vector Space Models of Semantics , 2010, J. Artif. Intell. Res..

[8]  Alessandro Lenci,et al.  How we BLESSed distributional semantic evaluation , 2011, GEMS.

[9]  Katrin Erk,et al.  Measuring semantic relatedness with vector space models and random walks , 2009, Graph-based Methods for Natural Language Processing.

[10]  John A Bullinaria,et al.  Extracting semantic representations from word co-occurrence statistics: stop-lists, stemming, and SVD , 2012, Behavior Research Methods.

[11]  Magnus Sahlgren,et al.  The Word-Space Model: using distributional analysis to represent syntagmatic and paradigmatic relations between words in high-dimensional vector spaces , 2006 .

[12]  Nathan Halko,et al.  Finding Structure with Randomness: Probabilistic Algorithms for Constructing Approximate Matrix Decompositions , 2009, SIAM Rev..

[13]  Ken McRae,et al.  People Use their Knowledge of Common Events to Understand Language, and Do So as Quickly as Possible , 2009, Lang. Linguistics Compass.

[14]  Zellig S. Harris,et al.  Distributional Structure , 1954 .

[15]  Alessandro Lenci,et al.  Distributional Memory: A General Framework for Corpus-Based Semantics , 2010, CL.

[16]  Mary Hare,et al.  Activating event knowledge , 2009, Cognition.

[17]  J. Bullinaria,et al.  Extracting semantic representations from word co-occurrence statistics: A computational study , 2007, Behavior research methods.

[18]  Scott McDonald,et al.  A Distributional Model of Semantic Context Effects in Lexical Processing , 2004, ACL.

[19]  Michael N Jones,et al.  Representing word meaning and order information in a composite holographic lexicon. , 2007, Psychological review.

[20]  Michael J Cortese,et al.  Predicting semantic priming at the item level , 2008 .

[21]  Mirella Lapata,et al.  Dependency-Based Construction of Semantic Space Models , 2007, CL.

[22]  J. Elman,et al.  A basis for generating expectancies for verbs from nouns , 2005, Memory & cognition.

[23]  Mark Steyvers,et al.  Topics in semantic representation. , 2007, Psychological review.

[24]  K. McRae,et al.  Integrating Verbs, Situation Schemas, and Thematic Role Concepts , 2001 .

[25]  R. Ratcliff,et al.  Spreading activation versus compound cue accounts of priming: mediated priming revisited. , 1992, Journal of experimental psychology. Learning, memory, and cognition.