Automatic Speech Recognition System for Home Appliances Control

In the present work we study the performance of a speech recognizer for the Greek language, in a smart-home environment. This recognizer operates in spoken interaction scenarios, where the users are able to control various home appliances. In contrast to command and control systems, in our application the users speak spontaneously, beyond the use of a standardized set of isolated commands. The operational performance was tested over various environmental conditions, for two different types of microphones. In all experiments, regardless of the difference in the word error rates obtained for different scenarios, a task completion rate of 100% was observed.

[1]  Matthias Denecke Rapid Prototyping for Spoken Dialogue Systems , 2002, COLING.

[2]  Kuansan Wang A plan-based dialog system with probabilistic inferences , 2000, INTERSPEECH.

[3]  Rohit Kumar,et al.  Conquestâ - An Open-Source Dialog System for Conferences , 2007, HLT-NAACL.

[4]  R. Bakis Continuous speech recognition via centisecond acoustic states , 1976 .

[5]  Li Deng,et al.  Mipad: a next generation PDA prototype , 2000, INTERSPEECH.

[6]  Sadaoki Furui,et al.  Designing a multimodal dialogue system for information retrieval , 1998, ICSLP.

[7]  Manny Rayner,et al.  Comparing grammar-based and robust approaches to speech understanding: a case study , 2001, INTERSPEECH.

[8]  Maxine Eskénazi,et al.  Non-Native Users in the Let’s Go!! Spoken Dialogue System: Dealing with Linguistic Mismatch , 2004, NAACL.

[9]  Shan Lu,et al.  Towards a Dialogue System Based on Recognition and Synthesis of Japanese Sign Language , 1997, Gesture Workshop.

[10]  Nelson Morgan,et al.  Double the trouble: handling noise and reverberation in far-field automatic speech recognition , 2002, INTERSPEECH.

[11]  Aarne Ranta,et al.  The TALK Grammar Library: an Integration of GF with TrindiKit , 2005 .

[12]  Eric Sanders,et al.  Speechdat multilingual speech databases for teleservices: across the finish line , 1999, EUROSPEECH.

[13]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[14]  Alexander I. Rudnicky,et al.  The RavenClaw dialog management framework: Architecture and systems , 2009, Comput. Speech Lang..

[15]  Nikos Fakotakis,et al.  Greek speech database for creation of voice driven teleservices , 1997, EUROSPEECH.

[16]  Holmer Hemsen Designing a multimodal dialogue system for mobile phones , 2003 .

[17]  Alexander I. Rudnicky,et al.  TeamTalk: A Platform for Multi-Human-Robot Dialog Research in Coherent Real and Virtual Spaces , 2007, AAAI.

[18]  Jennifer Chu-Carroll,et al.  MIMIC: An Adaptive Mixed Initiative Spoken Dialogue System for Information Queries , 2000, ANLP.

[19]  Marilyn A. Walker,et al.  MATCH: An Architecture for Multimodal Dialogue Systems , 2002, ACL.

[20]  Todor Ganchev,et al.  A practical, real-time speech-driven home automation front-end , 2005, IEEE Transactions on Consumer Electronics.

[21]  Wolfgang Wahlster,et al.  Smartkom: multimodal communication with a life- like character , 2001, INTERSPEECH.

[22]  Sebastian Möller,et al.  INSPIRE: Evaluation of a Smart-Home System for Infotainment Management and Device Control , 2004, LREC.