Integrating Pointing Gestures into a Spanish-spoken Dialog System for Conversational Service Robots

In this paper we present our work on the integration of human pointing gestures into a spoken dialog system in Spanish for conversational service robots. The dialog system is composed by a dialog manager, an interpreter that guides the spoken dialog and robot actions, in terms of user intentions and relevant environment stimuli associated to the current conversational situation. We demonstrate our approach by developing a tour–guide robot that is able to move around its environment, visually recognize informational posters, and explain sections of the poster selected by the user via pointing gestures. This robot also incorporates simple methods to qualify confidence in its visual outcomes, to inform about its internal state, and to start error–prevention dialogs whenever necessary. Our results show the reliability of the overall approach to model complex multimodal human–robot interactions.

[1]  L. Enrique Sucar,et al.  Markovito: A Flexible and General Service Robot , 2009 .

[2]  R. Dillmann,et al.  Using gesture and speech control for commanding a robot assistant , 2002, Proceedings. 11th IEEE International Workshop on Robot and Human Interactive Communication.

[3]  Luis Villaseñor Pineda,et al.  DIMEx100: A New Phonetic and Speech Corpus for Mexican Spanish , 2004, IBERAMIA.

[4]  Tetsunori Kobayashi,et al.  A conversational robot utilizing facial and body expressions , 2000, Smc 2000 conference proceedings. 2000 ieee international conference on systems, man and cybernetics. 'cybernetics evolving to systems, humans, organizations, and their complex interactions' (cat. no.0.

[5]  Jannik Fritsch,et al.  Modality integration and dialog management for a robotic assistant , 2005, INTERSPEECH.

[6]  Frédéric Lerasle,et al.  Mutual assistance between speech and vision for human-robot interaction , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[7]  Alexander H. Waibel,et al.  Enabling Multimodal Human–Robot Interaction for the Karlsruhe Humanoid Robot , 2007, IEEE Transactions on Robotics.