Generation and Evaluation of Communicative Robot Gesture

How is communicative gesture behavior in robots perceived by humans? Although gesture is crucial in social interaction, this research question is still largely unexplored in the field of social robotics. Thus, the main objective of the present work is to investigate how gestural machine behaviors can be used to design more natural communication in social robots. The chosen approach is twofold. Firstly, the technical challenges encountered when implementing a speech-gesture generation model on a robotic platform are tackled. We present a framework that enables the humanoid robot to flexibly produce synthetic speech and co-verbal hand and arm gestures at run-time, while not being limited to a predefined repertoire of motor actions. Secondly, the achieved flexibility in robot gesture is exploited in controlled experiments. To gain a deeper understanding of how communicative robot gesture might impact and shape human perception and evaluation of human-robot interaction, we conducted a between-subjects experimental study using the humanoid robot in a joint task scenario. We manipulated the non-verbal behaviors of the robot in three experimental conditions, so that it would refer to objects by utilizing either (1) unimodal (i.e., speech only) utterances, (2) congruent multimodal (i.e., semantically matching speech and gesture) or (3) incongruent multimodal (i.e., semantically non-matching speech and gesture) utterances. Our findings reveal that the robot is evaluated more positively when non-verbal behaviors such as hand and arm gestures are displayed along with speech, even if they do not semantically match the spoken utterance.

[1]  David P. Wilkins Adam Kendon (2004). Gesture: Visible action as utterance , 2006 .

[2]  Robert Dale,et al.  Building Natural Language Generation Systems: Figures , 2000 .

[3]  Stefan Kopp,et al.  A Knowledge-based Approach for Lifelike Gesture Animation , 2000, ECAI.

[4]  T. Mexia,et al.  Author ' s personal copy , 2009 .

[5]  Aude Billard,et al.  Learning of Gestures by Imitation in a Humanoid Robot , 2007 .

[6]  Nicolas Courty,et al.  Gesture in Human-Computer Interaction and Simulation , 2006 .

[7]  Autumn B. Hostetter,et al.  When do gestures communicate? A meta-analysis. , 2011, Psychological bulletin.

[8]  Takayuki Kanda,et al.  Natural deictic communication with humanoid robots , 2007, 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[9]  S. Goldin-Meadow,et al.  The role of gesture in communication and thinking , 1999, Trends in Cognitive Sciences.

[10]  Moonis Ali,et al.  Innovations in Applied Artificial Intelligence , 2005 .

[11]  H. Ishiguro,et al.  The thing that should not be: predictive coding and the uncanny valley in perceiving human and humanoid robot actions , 2011, Social cognitive and affective neuroscience.

[12]  Stefan Kopp,et al.  A friendly gesture: Investigating the effect of multimodal robot behavior in human-robot interaction , 2011, 2011 RO-MAN.

[13]  Stefan Kopp,et al.  Lifelike Gesture Synthesis and Timing for Conversational Agents , 2001, Gesture Workshop.

[14]  Takayuki Kanda,et al.  Providing route directions: Design of robot's utterance, gesture, and timing , 2009, 2009 4th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[15]  Marc Schröder,et al.  The German Text-to-Speech Synthesis System MARY: A Tool for Research, Development and Teaching , 2003, Int. J. Speech Technol..

[16]  Pengcheng Luo,et al.  Synchronized gesture and speech production for humanoid robots , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[17]  Stefan Kopp,et al.  Individualized Gesturing Outperforms Average Gesturing - Evaluating Gesture Production in Virtual Humans , 2010, IVA.

[18]  Stefan Kopp,et al.  Multimodal Communication from Multimodal Thinking - towards an Integrated Model of Speech and Gesture Production , 2008, Int. J. Semantic Comput..

[19]  Norihiro Hagita,et al.  Gesture Translation for Heterogeneous Robots , 2006, 2006 6th IEEE-RAS International Conference on Humanoid Robots.

[20]  Christian Heath,et al.  IEEE International Symposium on Robot and Human Interactive Communication , 2009 .

[21]  Autumn B. Hostetter,et al.  Visible embodiment: Gestures as simulated action , 2008, Psychonomic bulletin & review.

[22]  N. Lesh,et al.  The Role of Dialog in Human Robot Interaction , 2003 .

[23]  Ipke Wachsmuth,et al.  Gesture and Sign Language in Human-Computer Interaction , 1998, Lecture Notes in Computer Science.

[24]  Sven Behnke,et al.  Fritz - A Humanoid Communication Robot , 2007, RO-MAN 2007 - The 16th IEEE International Symposium on Robot and Human Interactive Communication.

[25]  K. Dautenhahn,et al.  Imitation and Social Learning in Robots, Humans and Animals: Behavioural, Social and Communicative Dimensions , 2009 .

[26]  Maurizio Mancini,et al.  Implementing Expressive Gesture Synthesis for Embodied Conversational Agents , 2005, Gesture Workshop.

[27]  T. Belpraeme,et al.  Imitation and Social Learning in Robots, Humans and Animals: Behavioural, Social and Communicative Dimensions , 2006 .

[28]  Stefan Kopp,et al.  The Effects of an Embodied Conversational Agent's Nonverbal Behavior on User's Evaluation and Behavioral Mimicry , 2007, IVA.

[29]  H. Ishiguro,et al.  The uncanny advantage of using androids in cognitive and social science research , 2006 .

[30]  Stefan Schaal,et al.  Robot Programming by Demonstration , 2009, Springer Handbook of Robotics.

[31]  D. McNeill Gesture and Thought , 2005 .

[32]  Stefan Kopp,et al.  Generating robot gesture using a virtual agent framework , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[33]  C. Creider Hand and Mind: What Gestures Reveal about Thought , 1994 .

[34]  Hiroshi Ishiguro,et al.  Development of an android robot for studying human-robot interaction , 2004 .

[35]  M. Matarić,et al.  Fixation behavior in observation and imitation of human movement. , 1998, Brain research. Cognitive brain research.

[36]  Michael Gienger,et al.  Task-oriented whole body motion for humanoid robots , 2005, 5th IEEE-RAS International Conference on Humanoid Robots, 2005..

[37]  Fokie Cnossen,et al.  Proceedings of the 15th IEEE International Symposium on Robot and Human Interactive Communication , 2006 .

[38]  Radoslaw Niewiadomski,et al.  Greta: an interactive expressive ECA system , 2009, AAMAS.

[39]  Stefan Kopp,et al.  Towards an integrated model of speech and gesture production for multi-modal robot behavior , 2010, 19th International Symposium in Robot and Human Interactive Communication.

[40]  J. Cassell,et al.  Embodied conversational agents , 2000 .

[41]  Christopher G. Atkeson,et al.  Adapting human motion for the control of a humanoid robot , 2002, Proceedings 2002 IEEE International Conference on Robotics and Automation (Cat. No.02CH37292).

[42]  Robert J. Wood,et al.  Towards a 3g crawling robot through the integration of microrobot technologies , 2006, Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006..

[43]  Stefan Kopp,et al.  Synthesizing multimodal utterances for conversational agents , 2004, Comput. Animat. Virtual Worlds.

[44]  K. Dautenhahn,et al.  The correspondence problem , 2002 .

[45]  Adam Kendon Gesture: Visible Action as Utterance: Gesture and referential meaning , 2004 .

[46]  Justine Cassell,et al.  Human conversation as a system framework: designing embodied conversational agents , 2001 .

[47]  Ehud Reiter,et al.  Book Reviews: Building Natural Language Generation Systems , 2000, CL.

[48]  María Malfaz,et al.  Multimodal Human-Robot Interaction Framework for a Personal Robot , 2006, ROMAN 2006 - The 15th IEEE International Symposium on Robot and Human Interactive Communication.

[49]  A. Takanishi,et al.  Various emotional expressions with emotion expression humanoid robot WE-4RII , 2004, IEEE Conference on Robotics and Automation, 2004. TExCRA Technical Exhibition Based..

[50]  A. Kendon Gesticulation and Speech: Two Aspects of the Process of Utterance , 1981 .

[51]  Heloir,et al.  The Uncanny Valley , 2019, The Animation Studies Reader.

[52]  Stefan Kopp,et al.  MURML: A Multimodal Utterance Representation Markup Language for Conversational Agents , 2002 .