Impact of sensory preferences of individuals with autism on the recognition of emotions expressed by two robots, an avatar, and a human

We design a personalized human-robot environment for social learning for individuals with autism spectrum disorders (ASD). In order to define an individual’s profile, we posit that the individual’s reliance on proprioceptive and kinematic visual cues should affect the way the individual suffering from ASD interacts with a social agent (human/robot/virtual agent). In this paper, we assess the potential link between recognition performances of body/facial expressions of emotion of increasing complexity, emotion recognition on platforms with different visual features (two mini-humanoid robots, a virtual agent, and a human), and proprioceptive and visual cues integration of an individual. First, we describe the design of the EMBODI-EMO database containing videos of controlled body/facial expressions of emotions from various platforms. We explain how we validated this database with typically developed (TD) individuals. Then, we investigate the relationship between emotion recognition and proprioceptive and visual profiles of TD individuals and individuals with ASD. For TD individuals, our results indicate a relationship between profiles and emotion recognition. As expected, we show that TD individuals that rely more heavily on visual cues yield better recognition scores. However, we found that TD individuals relying on proprioception have better recognition scores, going against our hypothesis. Finally, participants with ASD relying more heavily on proprioceptive cues have lower emotion recognition scores on all conditions than participants relying on visual cues.

[1]  Elizabeth S. Kim,et al.  Social Robots as Embedded Reinforcers of Social Behavior in Children with Autism , 2012, Journal of Autism and Developmental Disorders.

[2]  M. Silver,et al.  Evaluation of a New Computer Intervention to Teach People with Autism or Asperger Syndrome to Recognize and Predict Emotions in Others , 2001, Autism : the international journal of research and practice.

[3]  D. McCloskey,et al.  Proprioceptive Illusions Induced by Muscle Vibration: Contribution by Muscle Spindles to Perception? , 1972, Science.

[4]  S. Coates,et al.  Field Dependence—Independence, Social—Non-Social Play and Sex Differences in Preschool Children , 1975, Perceptual and motor skills.

[5]  S. Baron-Cohen,et al.  Enhancing Emotion Recognition in Children with Autism Spectrum Conditions: An Intervention Using Animated Vehicles with Real Emotional Faces , 2010, Journal of autism and developmental disorders.

[6]  Tony Charman,et al.  Infants with autism: an investigation of empathy, pretend play, joint attention, and imitation. , 1997 .

[7]  Mustafa Suphi Erden,et al.  Emotional Postures for the Humanoid-Robot Nao , 2013, Int. J. Soc. Robotics.

[8]  Lawrie S. McKay,et al.  Vision in autism spectrum disorders , 2009, Vision Research.

[9]  M. Terwogt,et al.  Attention to facial emotion expressions in children with autism , 2006, Autism : the international journal of research and practice.

[10]  K. Scherer,et al.  Emotion expression in body action and posture. , 2012, Emotion.

[11]  Angelo Cappello,et al.  Stabilometric parameters are affected by anthropometry and foot placement. , 2002, Clinical biomechanics.

[12]  J. Roll,et al.  Kinaesthetic role of muscle afferents in man, studied by tendon vibration and microneurography , 2004, Experimental Brain Research.

[13]  K. Scherer Expression of emotion in voice and music. , 1995, Journal of voice : official journal of the Voice Foundation.

[14]  D. Mestre,et al.  Brief Report: Postural Reactivity to Fast Visual Motion Differentiates Autistic from Children with Asperger Syndrome , 2002, Journal of autism and developmental disorders.

[15]  Laurent Mottron,et al.  Postural Hypo-Reactivity in Autism is Contingent on Development and Visual Environment: A Fully Immersive Virtual Reality Study , 2011, Journal of autism and developmental disorders.

[16]  Rajesh P. N. Rao,et al.  "Social" robots are psychological agents for infants: A test of gaze following , 2010, Neural Networks.

[17]  Adriana Tapus,et al.  Individuals with Autism: Analysis of the First Interaction with Nao Robot Based on Their Proprioceptive and Kinematic Profiles , 2015, RAAD.

[18]  A. Hamilton,et al.  Motor Abilities in Autism: A Review Using a Computational Context , 2013, Journal of autism and developmental disorders.

[19]  Minoru Asada,et al.  Autism simulator employing augmented reality: A prototype , 2014, 4th International Conference on Development and Learning and on Epigenetic Robotics.

[20]  B Isableu,et al.  We are most aware of our place in the world when about to fall , 2004, Current Biology.

[21]  Ning Tan,et al.  The Role of Body Postures in the Recognition of Emotions in Contextually Rich Scenarios , 2014, Int. J. Hum. Comput. Interact..

[22]  Céline Clavel,et al.  MARC: a framework that features emotion models for facial animation during human–computer interaction , 2013, Journal on Multimodal User Interfaces.

[23]  A. Tapus,et al.  Children with Autism Social Engagement in Interaction with Nao, an Imitative Robot - A Series of Single Case Experiments , 2012 .

[24]  Mary Hart,et al.  Autism/excel study , 2005, Assets '05.

[25]  Olivia N. Saracho,et al.  Matching Teachers' and Students' Cognitive Styles , 2003 .

[26]  Kerstin Dautenhahn,et al.  Collaborating with Kaspar: Using an autonomous humanoid robot to foster cooperative dyadic play among children with autism , 2010, 2010 10th IEEE-RAS International Conference on Humanoid Robots.

[27]  Nicolas Vuillerme,et al.  Differential integration of visual and kinaesthetic signals to upright stance , 2011, Experimental Brain Research.

[28]  Cristina P. Santos,et al.  Building a game scenario to encourage children with autism to recognize and label emotions using a humanoid robot , 2014, The 23rd IEEE International Symposium on Robot and Human Interactive Communication.

[29]  Y. Bar-Haim,et al.  Motor Function and Social Participation in Kindergarten Children , 2006 .

[30]  H. Wallbott Bodily expression of emotion , 1998 .

[31]  H. Meeren,et al.  Rapid perceptual integration of facial expression and emotional body language. , 2005, Proceedings of the National Academy of Sciences of the United States of America.

[32]  A. Bondy,et al.  The Picture Exchange Communication System , 1994, Seminars in speech and language.

[33]  D. McCloskey,et al.  The role of joint receptors in human kinaesthesia when intramuscular receptors cannot contribute. , 1987, The Journal of physiology.

[34]  David A. Ziegler,et al.  Localization of white matter volume increase in autism and developmental language disorder , 2004, Annals of neurology.

[35]  Reuven Kohen-Raz,et al.  Postural control in children with autism , 1992, Journal of autism and developmental disorders.

[36]  François Michaud,et al.  How wild is wild? A taxonomy to characterize the ‘wildness’ of child-robot interaction , 2010, Int. J. Soc. Robotics.

[37]  S. Baron-Cohen,et al.  Is There a "Language of the Eyes"? Evidence from Normal Adults, and Adults with Autism or Asperger Syndrome , 1997 .

[38]  R. Shadmehr,et al.  Representation of internal models of action in the autistic brain , 2009, Nature Neuroscience.

[39]  Wenhao Liu,et al.  Field dependence–independence and physical activity engagement among middle school students , 2009 .

[40]  Beatrice de Gelder,et al.  The Bodily Expressive Action Stimulus Test (BEAST). Construction and validation of a stimulus basis for measuring perception of whole body expression of emotions , 2011 .

[41]  B. Edin Cutaneous afferents provide information about knee joint movements in humans , 2001, The Journal of physiology.

[42]  Mark C. Coulson Attributing Emotion to Static Body Postures: Recognition Accuracy, Confusions, and Viewpoint Dependence , 2004 .

[43]  Hideki Kozima,et al.  Interactive robots for communication-care: a case-study in autism therapy , 2005, ROMAN 2005. IEEE International Workshop on Robot and Human Interactive Communication, 2005..

[44]  Tom Ziemke,et al.  Robot-assisted therapy for autism spectrum disorders with (partially) autonomous control: Challenges and outlook , 2012, Paladyn J. Behav. Robotics.

[45]  D. Burke,et al.  The responses of human muscle spindle endings to vibration of non‐contracting muscles. , 1976, The Journal of physiology.

[46]  D. Mestre,et al.  Postural effects of motion vision in young autistic children. , 1995 .

[47]  Chrystopher L. Nehaniv,et al.  Title of paper : KASPAR – A Minimally Expressive Humanoid Robot for Human-Robot Interaction Research , 2009 .

[48]  S. Baron-Cohen,et al.  The "Reading the Mind in the Eyes" Test revised version: a study with normal adults, and adults with Asperger syndrome or high-functioning autism. , 2001, Journal of child psychology and psychiatry, and allied disciplines.

[49]  G. Celani,et al.  The Understanding of the Emotional Meaning of Facial Expressions in People with Autism , 1999, Journal of autism and developmental disorders.

[50]  B. Scassellati,et al.  Robots for use in autism research. , 2012, Annual review of biomedical engineering.

[51]  D. Feil-Seifer,et al.  Defining socially assistive robotics , 2005, 9th International Conference on Rehabilitation Robotics, 2005. ICORR 2005..