Multimodal Integration of Emotional Signals from Voice, Body, and Context: Effects of (In)Congruence on Emotion Recognition and Attitudes Towards Robots

Humanoid social robots have an increasingly prominent place in today’s world. Their acceptance in social and emotional human–robot interaction (HRI) scenarios depends on their ability to convey well recognized and believable emotional expressions to their human users. In this article, we incorporate recent findings from psychology, neuroscience, human–computer interaction, and HRI, to examine how people recognize and respond to emotions displayed by the body and voice of humanoid robots, with a particular emphasis on the effects of incongruence. In a social HRI laboratory experiment, we investigated contextual incongruence (i.e., the conflict situation where a robot’s reaction is incongrous with the socio-emotional context of the interaction) and cross-modal incongruence (i.e., the conflict situation where an observer receives incongruous emotional information across the auditory (vocal prosody) and visual (whole-body expressions) modalities). Results showed that both contextual incongruence and cross-modal incongruence confused observers and decreased the likelihood that they accurately recognized the emotional expressions of the robot. This, in turn, gives the impression that the robot is unintelligent or unable to express “empathic” behaviour and leads to profoundly harmful effects on likability and believability. Our findings reinforce the need of proper design of emotional expressions for robots that use several channels to communicate their emotional states in a clear and effective way. We offer recommendations regarding design choices and discuss future research areas in the direction of multimodal HRI.

[1]  Reid G. Simmons,et al.  Laban head-motions convey robot state: A call for robot body language , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[2]  Dana Kulic,et al.  Measurement Instruments for the Anthropomorphism, Animacy, Likeability, Perceived Intelligence, and Perceived Safety of Robots , 2009, Int. J. Soc. Robotics.

[3]  Aryel Beck,et al.  Towards an Affect Space for robots to display emotional body language , 2010, 19th International Symposium in Robot and Human Interactive Communication.

[4]  Maja J. Mataric,et al.  Human Perception of Audio-Visual Synthetic Character Emotion Expression in the Presence of Ambiguous and Conflicting Information , 2009, IEEE Transactions on Multimedia.

[5]  Manfred Tscheligi,et al.  Can You Read My Face? , 2015, Int. J. Soc. Robotics.

[6]  Illah R. Nourbakhsh,et al.  A survey of socially interactive robots , 2003, Robotics Auton. Syst..

[7]  James M. Conrad,et al.  Human-robot collaboration: A survey , 2015, SoutheastCon 2015.

[8]  James A. Russell,et al.  Do facial expressions signal specific emotions? Judging emotion from the face in context , 1996 .

[9]  P. Marler,et al.  Communication Goes Multimodal , 1999, Science.

[10]  F. Gosselin,et al.  Audio-visual integration of emotion expression , 2008, Brain Research.

[11]  J. Vroomen,et al.  The perception of emotions by ear and by eye , 2000 .

[12]  R. Dolan,et al.  The Kuleshov Effect: the influence of contextual framing on emotional attributions. , 2006, Social cognitive and affective neuroscience.

[13]  Goldie Nejat,et al.  Recognizing Emotional Body Language Displayed by a Human-like Social Robot , 2014, International Journal of Social Robotics.

[14]  Ran R. Hassin,et al.  Angry, Disgusted, or Afraid? , 2008, Psychological science.

[15]  Mark C. Coulson Attributing Emotion to Static Body Postures: Recognition Accuracy, Confusions, and Viewpoint Dependence , 2004 .

[16]  Selma Sabanovic,et al.  Perceptions of Affective Expression in a minimalist robotic face , 2013, 2013 8th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[17]  Michael Erb,et al.  Audiovisual integration of emotional signals in voice and face: An event-related fMRI study , 2007, NeuroImage.

[18]  Maja Pantic,et al.  Automatic Analysis of Facial Expressions: The State of the Art , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[19]  C. Krumhansl,et al.  Cross-modal interactions in the perception of musical performance , 2006, Cognition.

[20]  Mitsuru Ishizuka,et al.  Evaluating Affective Feedback of the 3D Agent Max in a Competitive Cards Game , 2005, ACII.

[21]  H. Meeren,et al.  Rapid perceptual integration of facial expression and emotional body language. , 2005, Proceedings of the National Academy of Sciences of the United States of America.

[22]  Russell Beale,et al.  Psychological responses to simulated displays of mismatched emotional expressions , 2008, Interact. Comput..

[23]  Nikolaos Mavridis,et al.  A review of verbal and non-verbal human-robot interactive communication , 2014, Robotics Auton. Syst..

[24]  George P. Knight,et al.  Empathy-Related Responding and Cognition: A “Chicken and the Egg” Dilemma , 2014 .

[25]  Robert B. Welch,et al.  The interaction of vision and audition in two-dimensional space , 2015, Front. Neurosci..

[26]  Amanda J. C. Sharkey,et al.  It's not all written on the robot's face , 2012, Robotics Auton. Syst..

[27]  K. Scherer,et al.  In the eye of the beholder? Universality and cultural specificity in the expression and perception of emotion. , 2011, International journal of psychology : Journal international de psychologie.

[28]  Anthony G. Pipe,et al.  Perception of own and robot engagement in human-robot interactions and their dependence on robotics knowledge , 2014, Robotics Auton. Syst..

[29]  Andrea Kleinsmith,et al.  Affective Body Expression Perception and Recognition: A Survey , 2013, IEEE Transactions on Affective Computing.

[30]  Selma Sabanovic,et al.  Context congruency and robotic facial expressions: Do effects on human perceptions vary across culture? , 2014, The 23rd IEEE International Symposium on Robot and Human Interactive Communication.

[31]  Selma Sabanovic,et al.  Deriving Minimal Features for Human-Like Facial Expressions in Robotic Faces , 2014, International Journal of Social Robotics.

[32]  Ana Paiva,et al.  Metrics for Character Believability in Interactive Narrative , 2013, ICIDS.

[33]  P. Ekman An argument for basic emotions , 1992 .

[34]  Cynthia Breazeal,et al.  Emotion and sociable humanoid robots , 2003, Int. J. Hum. Comput. Stud..

[35]  Cristina P. Santos,et al.  Facial Expressions and Gestures to Convey Emotions with a Humanoid Robot , 2013, ICSR.

[36]  Pei-Luen Patrick Rau,et al.  A Cross-cultural Study: Effect of Robot Appearance and Task , 2010, Int. J. Soc. Robotics.

[37]  B. de Gelder,et al.  Body expressions influence recognition of emotions in the face and voice. , 2007, Emotion.

[38]  C. Nass,et al.  When a Talking-Face Computer Agent Is Half-Human and Half-Humanoid: Human Identity and Consistency Preference. , 2007 .

[39]  H. Wagner On measuring performance in category judgment studies of nonverbal behavior , 1993 .

[40]  Amir Aly,et al.  Towards an Interactive Human-Robot Relationship: Developing a Customized Robot Behavior to Human Profile. (Vers une relation Homme-Robot Interactive : développement d'un comportement du Robot adapté au Profil de l'Homme) , 2014 .

[41]  R. N. Spreng,et al.  The Toronto Empathy Questionnaire: Scale Development and Initial Validation of a Factor-Analytic Solution to Multiple Empathy Measures , 2009, Journal of personality assessment.

[42]  Emily S. Cross,et al.  The Perception of Emotion in Artificial Agents , 2018, IEEE Transactions on Cognitive and Developmental Systems.

[43]  Jennifer H. Pfeifer,et al.  Understanding emotions in others: mirror neuron dysfunction in children with autism spectrum disorders , 2006, Nature Neuroscience.

[44]  Adriana Tapus,et al.  Multimodal adapted robot behavior synthesis within a narrative human-robot interaction , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[45]  Nadia Bianchi-Berthouze,et al.  Modeling human affective postures: an information theoretic characterization of posture features , 2004, Comput. Animat. Virtual Worlds.

[46]  Atsuo Takanishi,et al.  Whole body emotion expressions for KOBIAN humanoid robot — preliminary experiments with different Emotional patterns — , 2009, RO-MAN 2009 - The 18th IEEE International Symposium on Robot and Human Interactive Communication.

[47]  Candace L. Sidner,et al.  Explorations in engagement for humans and robots , 2005, Artif. Intell..

[48]  Majid Nili Ahmadabadi,et al.  Multisensory Perception of Contradictory Information in an Environment of Varying Reliability: Evidence for Conscious Perception and Optimal Causal Inference , 2017, Scientific Reports.

[49]  Gilles Pourtois,et al.  Multisensory perception of emotion, its time course, and its neural basis , 2004 .

[50]  Lola Cañamero,et al.  I show you how I like you - can you read it in my face? [robotics] , 2001, IEEE Trans. Syst. Man Cybern. Part A.

[51]  M. D. Meijer The contribution of general features of body movement to the attribution of emotions , 1989 .

[52]  P. Bertelson,et al.  Multisensory integration, perception and ecological validity , 2003, Trends in Cognitive Sciences.

[53]  Stefan Kopp,et al.  A friendly gesture: Investigating the effect of multimodal robot behavior in human-robot interaction , 2011, 2011 RO-MAN.

[54]  Shlomo Hareli,et al.  What's Social About Social Emotions? , 2008 .

[55]  Sylvia D. Kreibig,et al.  Autonomic nervous system activity in emotion: A review , 2010, Biological Psychology.

[56]  Hillary Anger Elfenbein,et al.  On the universality and cultural specificity of emotion recognition: a meta-analysis. , 2002, Psychological bulletin.

[57]  J. Gross,et al.  Emotion elicitation using films , 1995 .

[58]  Beatrice de Gelder,et al.  Emotional Voice and Emotional Body Postures Influence Each Other Independently of Visual Awareness , 2011, PloS one.

[59]  D. De Rossi,et al.  Can a Humanoid Face be Expressive? A Psychophysiological Investigation , 2015, Front. Bioeng. Biotechnol..

[60]  B. Wielinga,et al.  Effects of (in)accurate empathy and situational valence on attitudes towards robots , 2010, IEEE/ACM International Conference on Human-Robot Interaction.

[61]  Claudia Wagner,et al.  Anthropomorphic inferences from emotional nonverbal cues: A case study , 2010, 19th International Symposium in Robot and Human Interactive Communication.

[62]  B. Mesquita,et al.  Context in Emotion Perception , 2011 .

[63]  Katarzyna Wac,et al.  Designing Emotionally Expressive Robots: A Comparative Study on the Perception of Communication Modalities , 2017, HAI.

[64]  Tony Belpaeme,et al.  Non-Linguistic Utterances Should be Used Alongside Language, Rather than on their Own or as a Replacement , 2014, 2014 9th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[65]  Cindy L. Bethel,et al.  A Survey of Using Vocal Prosody to Convey Emotion in Robot Speech , 2016, Int. J. Soc. Robotics.

[66]  Aaron Powers,et al.  Matching robot appearance and behavior to tasks to improve human-robot cooperation , 2003, The 12th IEEE International Workshop on Robot and Human Interactive Communication, 2003. Proceedings. ROMAN 2003..

[67]  D. C. Howell Statistical Methods for Psychology , 1987 .

[68]  Jean-Claude Martin,et al.  Combining Facial and Postural Expressions of Emotions in a Virtual Character , 2009, IVA.

[69]  J. Russell A circumplex model of affect. , 1980 .