Towards Computational Guessing of Unknown Word Meanings: The Ontological Semantic Approach

Towards Computational Guessing of Unknown Word Meanings: The Ontological Semantic Approach Julia M. Taylor (jtaylor1@purdue.edu) CERIAS, Purdue University & RiverGlass, Inc West Lafayette, IN 47907 & Champaign, IL 61820 Victor Raskin (vraskin@purdue.edu) Linguistics & CERIAS, Purdue University West Lafayette, IN 47907 Christian F. Hempelmann (chempelm@purdue.edu) Linguistics, Purdue University & RiverGlass, Inc West Lafayette, IN 47907 & Champaign, IL 61820 Abstract The paper describes a computational approach for guessing the meanings of previously unaccounted words in an implemented system for natural language processing. Interested in comparing the results to what is known about human guessing, it reviews a largely educational approach, partially based on cognitive psychology, to teaching humans, mostly children, to acquire new vocabulary from contextual clues, as well as the lexicographic efforts to account for neologisms. It then goes over the previous NLP efforts in processing new words and establishes the difference—mostly, much richer semantic resources—of the proposed approach. Finally, the results of a computer experiment that guesses the meaning of a non-existent word, placed as the direct object of 100 randomly selected verbs, from the known meanings of these verbs, with methods of the ontological semantics technology, are presented and discussed. While the results are promising percentage-wise, ways to improve them within the approach are briefly outlined. Keywords: guessing word meaning, natural language understanding, ontological semantic technology Unknown Words in Text Along with ambiguity, unattested input is one of the major problems for natural language processing systems. An NLP system is robust only if it can deal with unknown words. Yet, to deal with such words only makes sense when the rest of the sentence is understood. We take an approach here similar to that of a human learner that encounters an unfamiliar word and is able to approximate its meaning based on the rest of the sentence or its subsequent usages in other sentences. There are some suggested strategies in the human acquisition and understanding of unknown words. Some cases stand out as easy and almost self-explanatory. One of these cases is when a word is immediately explained. Such an explanation may be introduced by a that is phrase (To lose weight, one may have to follow a diet, that is, to limit the amount of food and to avoid eating certain foods.), or by apposition (Computers programs follow algorithms, ordered lists of instructions to perform.), or by examples (The earliest records of felines, for example, cats, tigers, lions, or leopards, are from millions of years ago.), or by providing the presumably known opposites for comparison through words like but, rather then, not (It is frigid outside, rather than warm and comfortable like yesterday.). Both in the case of human acquisition of new vocabulary and the machine attempt at guessing its meaning, these somewhat trivial instances, where the meaning of a new word is immediately explained, either by giving its definition or by examples, present no particular interest for us here. Besides, such cases are rather rare in regular expository texts because most writers do not bother to allow for vocabulary deficiency with regards to words with which they are well familiar themselves. Thus, it is the non-trivial cases, those without an attached explanation or description, that it is necessary to address when one is interested in designing a computer system for natural language understanding. On the other side of the spectrum lie words that can only be guessed through their functional description, not necessarily following the first use of an unknown word. These functional descriptions should be gathered throughout the document, or a number of documents, narrowing the original functional description, if necessary, or supplying other facets of it. For example, They used a Tim-Tim to navigate their way to the cabin on the lake. It took them almost half a day. They hadn’t checked if the maps had been recently updated on the device, and spent hours looking for roads that no longer existed. From the clues in the first sentence, Tim-Tim can be understood as a navigation instrument (including an atlas or a map) through an inverse function of the instrument of navigation. Since no other devices are mentioned, this navigation instrument can be considered the device from the third sentence whose maps can be periodically updated. It is essential, therefore, in situations of dispersed clues that co-reference (or antecedence) be established correctly—in this case, between device and Tim-Tim. Towards the middle of the spectrum are the cases where the description may immediately follow the first use of the word but without being helpfully triggered by phrases like for example or that is (He was rather taciturn. He didn’t like

[1]  Angelika Rieder 11 ( 1 & 2 ) 53 A cognitive view of incidental vocabulary acquisition : From text meaning to word meaning ? , 2002 .

[2]  E. Wisniewski Construal and Similarity in Conceptual Combination , 1996 .

[3]  Claire Cardie,et al.  A Case-Based Approach to Knowledge Acquisition for Domain-Specific Sentence Analysis , 1993, AAAI.

[4]  James R. Curran,et al.  Supersense Tagging of Unknown Nouns Using Semantic Similarity , 2005, ACL.

[5]  Christopher D. Manning,et al.  Classifying Unknown Proper Noun Phrases Without Context , 2002 .

[6]  Richard M. Schwartz,et al.  An Algorithm that Learns What's in a Name , 1999, Machine Learning.

[7]  Yugo Murawaki,et al.  Online Acquisition of Japanese Unknown Morphemes using Morphological Constraints , 2008, EMNLP.

[8]  J. Simpson,et al.  Neologism: The Long View , 2012 .

[9]  Victor Raskin,et al.  Accessing and manipulating meaning of textual and data information for information assurance and security and intelligence information , 2008, CSIIRW '08.

[10]  Yorick Wilks,et al.  Lexical Tuning , 2002, CICLing.

[11]  Richard K. Wagner,et al.  Vocabulary Acquisition: Implications for Reading Comprehension. , 2006 .

[12]  Marcia K. Johnson,et al.  Contextual prerequisites for understanding: Some investigations of comprehension and recall , 1972 .

[13]  E. Wisniewski Similarity, alignment, and conceptual combination: Comment on Estes and Glucksberg , 2000, Memory & cognition.

[14]  John Algeo,et al.  Where Do All the New Words Come from , 1980 .

[15]  Mary O'Neill,et al.  A systematic approach to the selection of neologisms for inclusion in a large monolingual dictionary , 2008 .

[16]  E. Wisniewski When concepts combine , 1997, Psychonomic bulletin & review.

[17]  Margaret G. McKeown,et al.  Vocabulary Development: All Contexts Are Not Created Equal , 2008 .

[18]  I. Nation How Large a Vocabulary Is Needed for Reading and Listening? , 2006 .

[19]  P. Johnson-Laird Mental models , 1989 .

[20]  Margaret G. McKeown,et al.  The Acquisition of Word Meaning from Context by Children of High and Low Ability. Outstanding Dissertation Monograph 1985. , 1985 .

[21]  Yorick Wilks,et al.  Making Preferences More Active , 1978, Artif. Intell..

[22]  Richard Granger,et al.  FOUL-UP: A Program that Figures Out Meanings of Words from Context , 1977, IJCAI.

[23]  Yoram Singer,et al.  Unsupervised Models for Named Entity Classification , 1999, EMNLP.

[24]  William E. Nagy,et al.  Learning Word Meanings From Context During Normal Reading , 1987 .

[25]  Christian Hempelmann,et al.  On an Automatic Acquisition Toolbox for Ontologies and Lexicons in Ontological Semantics , 2010, International Conference on Artificial Intelligence.

[26]  Antal van den Bosch,et al.  Integrating Seed Names and ngrams for a Named Entity List and Classifier , 2000, LREC.

[27]  Steven L. Lytinen,et al.  The Ups and Downs of Lexical Acquisition , 1994, AAAI.

[28]  Satoshi Sekine,et al.  A survey of named entity recognition and classification , 2007 .

[29]  Animesh Mukherjee,et al.  Investigation and modeling of the structure of texting language , 2007, International Journal of Document Analysis and Recognition (IJDAR).

[30]  Edward J. Wisniewski,et al.  Conceptual combination: Possibilities and esthetics. , 1997 .

[31]  A. Lehrer Understanding trendy neologisms , 2003 .

[32]  Earl F. Rankin,et al.  Reaction of Intermediate Grade Children to Contextual Clues , 1969 .

[33]  Joan P. Gipe,et al.  Investigating Techniques for Teaching Word Meanings , 1978 .

[34]  Patricia A. Herman,et al.  Learning Words from Context. , 1985 .

[35]  Christian Hempelmann,et al.  Application-Guided Ontological Engineering , 2010, IC-AI.

[36]  David Yarowsky,et al.  Language Independent Named Entity Recognition Combining Morphological and Contextual Evidence , 1999, EMNLP.

[37]  David K. Barnhart A Calculus for New Words , 2012 .

[38]  John Algeo,et al.  Fifty years among the new words : a dictionary of neologisms, 1941-1991 , 1993 .

[39]  C. Paul Cook,et al.  Exploiting Linguistic Knowledge to Infer Properties of Neologisms , 2011 .

[40]  Wilson L. Taylor,et al.  “Cloze Procedure”: A New Tool for Measuring Readability , 1953 .

[41]  Massimiliano Ciaramita Boosting automatic lexical acquisition with morphological information , 2002, ACL 2002.

[42]  Jesse T Sheidlower Principles for the Inclusion of New Words in College Dictionaries , 1995 .

[43]  David A. Campbell,et al.  A technique for semantic classification of unknown words using UMLS resources , 1999, AMIA.

[44]  John Algeo Desuetude among New English Words , 1993 .

[45]  Massimiliano Ciaramita,et al.  Supersense Tagging of Unknown Nouns in WordNet , 2003, EMNLP.

[46]  Richard M. Schwartz,et al.  Coping with Ambiguity and Unknown Words through Probabilistic Models , 1993, CL.

[47]  J. Algeo Blends, a Structural and Systemic View , 1977 .

[48]  Andrei Mikheev,et al.  Automatic Rule Induction for Unknown-Word Guessing , 1997, CL.

[49]  W. Kintsch,et al.  Strategies of discourse comprehension , 1983 .

[50]  Elinore Kress Schatz,et al.  Context Clues Are Unreliable Predictors of Word Meanings. , 1986 .