Mobile spoken dialogue system using parser dependencies and ontology

In this paper, we propose a mobile spoken dialogue system with a new spoken dialogue understanding architecture (SLU). This new SLU module combines an ontology and a dependency graph to do semantic analysis. The turn analysis algorithm integrated in the SLU module uses, at each turn of the dialogue, the dependencies generated by Stanford parser and a domain ontology to analyze the sentence and to extract user’s intention and slots values (i.e. user dialogue acts, concepts and their values). The SLU module maps the sentence into a structure. The dialogue manager receives this mapping structure from the SLU module and chooses the action to be taken by the system. The mobile spoken dialogue system was developed as a remote system on a mobile phone. It utilizes Google server for recognition and Google text to speech for speech synthesis provided with Android system. Dialogue understanding, dialogue managing and text generating modules reside on a remote computer. Ten users have tested the first version of our system and a score of 3.6 on Likert scale was obtained.

[1]  Gokhan Tur,et al.  Spoken Language Understanding: Systems for Extracting Semantic Information from Speech , 2011 .

[2]  Amal Zouaq,et al.  Evaluating the Generation of Domain Ontologies in the Knowledge Puzzle Project , 2009, IEEE Transactions on Knowledge and Data Engineering.

[3]  J. Gabriel Amores,et al.  Integrating OWL Ontologies with a Dialogue Manager , 2006, Proces. del Leng. Natural.

[4]  Alexander I. Rudnicky,et al.  Ravenclaw: dialog management using hierarchical task decomposition and an expectation agenda , 2003, INTERSPEECH.

[5]  Giuseppe Riccardi,et al.  What's in an ontology for spoken language understanding , 2009, INTERSPEECH.

[6]  Oliver Lemon,et al.  DIPPER: Description and Formalisation of an Information-State Update Dialogue System Architecture , 2003, SIGDIAL Workshop.

[7]  Hui Ye,et al.  The Hidden Information State Approach to Dialog Management , 2007, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07.

[8]  Gary Geunbae Lee,et al.  Recent Approaches to Dialog Management for Spoken Dialog Systems , 2010, J. Comput. Sci. Eng..

[9]  Alexander I. Rudnicky,et al.  Olympus: an open-source framework for conversational spoken language interface research , 2007, HLT-NAACL 2007.

[10]  James F. Allen,et al.  Integrating linguistic and domain knowledge for spoken dialogue systems in multiple domains , 2003, IJCAI 2003.

[11]  Alex Acero,et al.  Discriminative models for spoken language understanding , 2006, INTERSPEECH.

[12]  Giuseppe Riccardi,et al.  Generative and discriminative algorithms for spoken language understanding , 2007, INTERSPEECH.

[13]  Marilyn A. Walker,et al.  MATCH: An Architecture for Multimodal Dialogue Systems , 2002, ACL.

[14]  Thomas R. Gruber,et al.  A translation approach to portable ontology specifications , 1993, Knowl. Acquis..

[15]  James H. Martin,et al.  Speech and language processing: an introduction to natural language processing, computational linguistics, and speech recognition, 2nd Edition , 2000, Prentice Hall series in artificial intelligence.

[16]  Yarden Katz,et al.  Pellet: A practical OWL-DL reasoner , 2007, J. Web Semant..

[17]  Christopher D. Manning,et al.  The Stanford Typed Dependencies Representation , 2008, CF+CDPE@COLING.