Expressive facial speech synthesis on a robotic platform

This paper presents our expressive facial speech synthesis system Eface, for a social or service robot. Eface aims at enabling a robot to deliver information clearly with empathetic speech and an expressive virtual face. The empathetic speech is built on the Festival speech synthesis system and provides robots the capability to speak with different voices and emotions. Two versions of a virtual face have been implemented to display the robot's expressions. One with just over 100 polygons has a lower hardware requirement but looks less natural. The other has over 1000 polygons; it looks realistic, but costs more CPU resource and requires better video hardware. The whole system is incorporated into the popular open source robot interface Player, which makes client programs easy to write and debug. Also, it is convenient to use the same system with different robot platforms. We have implemented this system on a physical robot and tested it with a robotic nurse assistant scenario.

[1]  Fernando Pereira,et al.  MPEG-4 facial animation technology: survey, implementation, and results , 1999, IEEE Trans. Circuits Syst. Video Technol..

[2]  Richard T. Vaughan,et al.  The Player/Stage Project: Tools for Multi-Robot and Distributed Sensor Systems , 2003 .

[3]  Koray Balci Xface: MPEG-4 based open source toolkit for 3D Facial Animation , 2004, AVI.

[4]  Kim E. A. Silverman,et al.  Vocal cues to speaker affect: testing two models , 1984 .

[5]  Cynthia Breazeal,et al.  Designing sociable robots , 2002 .

[6]  Bruce MacDonald,et al.  Towards Expressive Speech Synthesis in English on a Robotic Platform , 2006 .

[7]  Paul Taylor,et al.  The architecture of the Festival speech synthesis system , 1998, SSW.

[8]  Prem Kalra,et al.  Face to virtual face , 1998, Proc. IEEE.

[9]  Ahmet M. Kondoz,et al.  Automatic Single View-Based 3-D Face Synthesis for Unsupervised Multimedia Applications , 2008, IEEE Transactions on Circuits and Systems for Video Technology.

[10]  Joelle Pineau,et al.  Towards robotic assistants in nursing homes: Challenges and results , 2003, Robotics Auton. Syst..

[11]  Cynthia Breazeal,et al.  Recognition of Affective Communicative Intent in Robot-Directed Speech , 2002, Auton. Robots.

[12]  Francis J. Hill Computer Graphics Using Open GL , 1990 .

[13]  Mari Ostendorf,et al.  TOBI: a standard for labeling English prosody , 1992, ICSLP.

[14]  Illah R. Nourbakhsh,et al.  An Affective Mobile Robot Educator with a Full-Time Job , 1999, Artif. Intell..

[15]  Brian Scassellati,et al.  How to build robots that make friends and influence people , 1999, Proceedings 1999 IEEE/RSJ International Conference on Intelligent Robots and Systems. Human and Environment Friendly Robots with High Intelligence and Emotional Quotients (Cat. No.99CH36289).

[16]  Wolfram Burgard,et al.  MINERVA: a second-generation museum tour-guide robot , 1999, Proceedings 1999 IEEE International Conference on Robotics and Automation (Cat. No.99CH36288C).

[17]  Jörgen Ahlberg,et al.  CANDIDE-3 - An Updated Parameterised Face , 2001 .

[18]  D. Fox,et al.  Towards Personal Service Robots for the Elderly , 1999 .

[19]  F. Hara,et al.  Use of face robot for human-computer communication , 1995, 1995 IEEE International Conference on Systems, Man and Cybernetics. Intelligent Systems for the 21st Century.

[20]  Christoph Bartneck,et al.  Interacting with an embodied emotional character , 2003, DPPI '03.

[21]  Franck Davoine,et al.  Expressive face recognition and synthesis , 2003, 2003 Conference on Computer Vision and Pattern Recognition Workshop.