Generating Example Contexts to Illustrate a Target Word Sense

Learning a vocabulary word requires seeing it in multiple informative contexts. We describe a system to generate such contexts for a given word sense. Rather than attempt to do word sense disambiguation on example contexts already generated or selected from a corpus, we compile information about the word sense into the context generation process. To evaluate the sense-appropriateness of the generated contexts compared to WordNet examples, three human judges chose which word sense(s) fit each example, blind to its source and intended sense. On average, one judge rated the generated examples as sense-appropriate, compared to two judges for the WordNet examples. Although the system's precision was only half of WordNet's, its recall was actually higher than WordNet's, thanks to covering many senses for which WordNet lacks examples.

[1]  Rada Mihalcea,et al.  Bootstrapping Large Sense Tagged Corpora , 2002, LREC.

[2]  Kees de Glopper,et al.  Deriving Word Meaning from Written Context: A Multicomponential Skill , 2001 .

[3]  William E. Nagy,et al.  Learning Word Meanings From Context During Normal Reading , 1987 .

[4]  Liu Liu,et al.  Automated generation of example contexts for helping children learn vocabulary , 2009, SLaTE.

[5]  Patricia A. Herman,et al.  Incidental acquisition of word meaning from expositions with varied text features , 1987 .

[6]  Vladimir M. Sloutsky,et al.  Learning Words from Context , 2008 .

[7]  David Yarowsky,et al.  Unsupervised Word Sense Disambiguation Rivaling Supervised Methods , 1995, ACL.

[8]  Eneko Agirre,et al.  On the Use of Automatically Acquired Examples for All-Nouns Word Sense Disambiguation , 2008, J. Artif. Intell. Res..

[9]  Liu Liu,et al.  Generating example contexts to help children learn word meaning† , 2012, Natural Language Engineering.

[10]  Ted Pedersen,et al.  Name Discrimination and Email Clustering using Unsupervised Clustering and Labeling of Similar Contexts , 2005, IICAI.

[11]  Steven A. Stahl,et al.  Partial Word Knowledge and Vocabulary Growth during Reading Comprehension , 1997 .

[12]  R. U S L A N M I T K O V,et al.  A computer-aided environment for generating multiple-choice test items , 2005 .

[13]  Margaret G. McKeown,et al.  Bringing Words to Life: Robust Vocabulary Instruction , 2002 .

[14]  David M. Blei,et al.  Probabilistic topic models , 2012, Commun. ACM.

[15]  Michael I. Jordan,et al.  Latent Dirichlet Allocation , 2001, J. Mach. Learn. Res..

[16]  Elinore Kress Schatz,et al.  Context Clues Are Unreliable Predictors of Word Meanings. , 1986 .

[17]  Le An Ha,et al.  A computer-aided environment for generating multiple-choice test items , 2006, Natural Language Engineering.

[18]  Suzanne M. Adlof,et al.  Learning Vocabulary through Reading , 2006 .

[19]  Alexander Yates,et al.  Extracting Glosses to Disambiguate Word Senses , 2010, HLT-NAACL.

[20]  Maxine Eskénazi,et al.  An Application of Latent Semantic Analysis to Word Sense Discrimination for Words with Related and Unrelated Meanings , 2009, BEA@NAACL.

[21]  J. Fleiss,et al.  Intraclass correlations: uses in assessing rater reliability. , 1979, Psychological bulletin.

[22]  Roberto Navigli,et al.  SemEval-2007 Task 07: Coarse-Grained English All-Words Task , 2007, Fourth International Workshop on Semantic Evaluations (SemEval-2007).

[23]  Steven A. Stahl,et al.  Teaching Children to Learn Word Meanings from Context: A Synthesis and Some Questions , 1998 .

[24]  Ramesh Nallapati,et al.  Labeled LDA: A supervised topic model for credit attribution in multi-labeled corpora , 2009, EMNLP.