Introduction to the Special Issue on Spoken Language Understanding in Conversational Systems

Understanding spoken language is about extracting the meaning from speech utterances. Although there continues to be endless debates in linguistics, philosophy, psychology, and neuroscience on what constitutes the meaning of a natural language utterance (Jackendoff, 2002), for the purpose of human– computer interactive systems, ‘‘meaning’’ is regarded as a representation that can be executed by an interpreter in order to change the state of the system. In such systems, understanding spoken language involves automatic speech recognition (ASR) and spoken language understanding (SLU)—a transduction of the recognition result to an interpretable representation. There are a number of aspects to SLU that makes this a challenging task. The most challenging of them all is the issue of meaning representation. Human language expresses meaning through a variety of surface forms, for example, prosody, lexical choice, syntax. The same meaning can be expressed in many different surface forms and also the same surface form can express many different meanings. These aspects are further accentuated in conversational systems where the dialog context plays a significant role in the meaning of an utterance. Designing a representation that captures this rich expressivity in full generality is a daunting task. So, in order to build practical systems, meaning representations tend to be crafted based on the desired capabilities

[1]  Gökhan Tür,et al.  Beyond ASR 1-best: Using word confusion networks in spoken language understanding , 2006, Comput. Speech Lang..

[2]  Rohit Prasad,et al.  Speech-enabled natural language call routing: BBN call director , 2002, INTERSPEECH.

[3]  Douglas E. Appelt,et al.  GEMINI: A Natural Language System for Spoken-Language Understanding , 1993, ACL.

[4]  Jean-Luc Gauvain,et al.  User evaluation of the MASK kiosk , 1998, Speech Commun..

[5]  Gregory A. Sanders,et al.  DARPA communicator: cross-system results for the 2001 evaluation , 2002, INTERSPEECH.

[6]  Giuseppe Riccardi,et al.  How may I help you? , 1997, Speech Commun..

[7]  Bob Carpenter,et al.  Vector-based Natural Language Call Routing , 1999, Comput. Linguistics.

[8]  Gökhan Tür,et al.  The AT&T spoken language understanding system , 2006, IEEE Transactions on Audio, Speech, and Language Processing.

[9]  Chin-Hui Lee,et al.  Automatic recognition of keywords in unconstrained speech using hidden Markov models , 1990, IEEE Trans. Acoust. Speech Signal Process..

[10]  Lou Boves,et al.  Overview of the ARISE project , 1999, EUROSPEECH.

[11]  Marilyn A. Walker,et al.  MATCH: An Architecture for Multimodal Dialogue Systems , 2002, ACL.

[12]  Wayne H. Ward,et al.  Recent Improvements in the CMU Spoken Language Understanding System , 1994, HLT.

[13]  Z. Harris,et al.  Foundations of Language , 1940 .

[14]  Marilyn A. Walker,et al.  PARADISE: A Framework for Evaluating Spoken Dialogue Agents , 1997, ACL.

[15]  Chung Hee Hwang,et al.  The TRAINS project: a case study in building a conversational planning agent , 1994, J. Exp. Theor. Artif. Intell..

[16]  P. J. Price,et al.  Evaluation of Spoken Language Systems: the ATIS Domain , 1990, HLT.

[17]  Stephanie Seneff,et al.  TINA: A Natural Language System for Spoken Language Applications , 1992, Comput. Linguistics.

[18]  Chin-Hui Lee,et al.  A speech understanding system based on statistical representation of semantics , 1992, [Proceedings] ICASSP-92: 1992 IEEE International Conference on Acoustics, Speech, and Signal Processing.

[19]  Richard M. Schwartz,et al.  Hidden Understanding Models of Natural Language , 1994, ACL.

[20]  Jeremy Peckham,et al.  Speech Understanding and Dialogue over the telephone: an overview of the ESPRIT SUNDIAL project. , 1991, HLT.

[21]  Lou Boves,et al.  A spoken dialog system for the Dutch public transport information service , 1997, Int. J. Speech Technol..