Smile Like You Mean It: Driving Animatronic Robotic Face with Learned Models

Ability to generate intelligent and generalizable facial expressions is essential for building human-like social robots. At present, progress in this field is hindered by the fact that each facial expression needs to be programmed by humans. In order to adapt robot behavior in real time to different situations that arise when interacting with human subjects, robots need to be able to train themselves without requiring human labels, as well as make fast action decisions and generalize the acquired knowledge to diverse and new contexts. We addressed this challenge by designing a physical animatronic robotic face with soft skin and by developing a vision-based self-supervised learning framework for facial mimicry. Our algorithm does not require any knowledge of the robot's kinematic model, camera calibration or predefined expression set. By decomposing the learning process into a generative model and an inverse model, our framework can be trained using a single motor babbling dataset. Comprehensive evaluations show that our method enables accurate and diverse face mimicry across diverse human subjects.

[1]  Pieter Abbeel,et al.  An Algorithmic Perspective on Imitation Learning , 2018, Found. Trends Robotics.

[2]  R. Plutchik Emotions : a general psychoevolutionary theory , 1984 .

[3]  Byron Boots,et al.  Agile Off-Road Autonomous Driving Using End-to-End Deep Imitation Learning , 2017, ArXiv.

[4]  Jason Yosinski,et al.  An Intriguing Failing of Convolutional Neural Networks and the CoordConv Solution , 2018, NeurIPS.

[5]  Tali Dekel,et al.  SinGAN: Learning a Generative Model From a Single Natural Image , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[6]  J. Piaget Play, dreams and imitation in childhood , 1951 .

[7]  Trevor Darrell,et al.  Fully Convolutional Networks for Semantic Segmentation , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[8]  Chyi-Yeu Lin,et al.  Humanoid Head Face Mechanism with Expandable Facial Expressions , 2016 .

[9]  Jun Morimoto,et al.  Learning from demonstration and adaptation of biped locomotion , 2004, Robotics Auton. Syst..

[10]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[11]  Alexey Dosovitskiy,et al.  End-to-End Driving Via Conditional Imitation Learning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[12]  N. Reissland Neonatal Imitation in the First Hour of Life : Observations in Rural Nepal , 2001 .

[13]  Carme Torras,et al.  A robot learning from demonstration framework to perform force-based manipulation tasks , 2013, Intelligent Service Robotics.

[14]  Hod Lipson,et al.  Visual behavior modelling for robotic theory of mind , 2021, Scientific reports.

[15]  Nicu Sebe,et al.  Animating Arbitrary Objects via Deep Motion Transfer , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[16]  T. Tsuji,et al.  Development of the Face Robot SAYA for Rich Facial Expressions , 2006, 2006 SICE-ICASE International Joint Conference.

[17]  Dong-Wook Lee,et al.  Designing of android head system by applying facial muscle mechanism of humans , 2012, 2012 12th IEEE-RAS International Conference on Humanoid Robots (Humanoids 2012).

[18]  David Loza,et al.  Application of the FACS in the design and construction of a mechatronic head with realistic appearance , 2013 .

[19]  Zhong Huang,et al.  Automatic Facial Expression Learning Method Based on Humanoid Robot XIN-REN , 2016, IEEE Transactions on Human-Machine Systems.

[20]  Hao Li,et al.  paGAN: real-time avatars using dynamic textures , 2019, ACM Trans. Graph..

[21]  Minoru Asada,et al.  Identification and Evaluation of the Face System of a Child Android Robot Affetto for Surface Motion Design , 2018, Front. Robot. AI.

[22]  Hiroshi Kobayashi,et al.  Dynamic display of facial expressions on the face robot made by using a life mask , 2008, Humanoids 2008 - 8th IEEE-RAS International Conference on Humanoid Robots.

[23]  Jochen J. Steil,et al.  Interactive imitation learning of object movement skills , 2011, Autonomous Robots.

[24]  Atsuo Takanishi,et al.  Mechanical design of emotion expression humanoid robot we-4rii , 2006 .

[25]  Ken Goldberg,et al.  Deep Imitation Learning for Complex Manipulation Tasks from Virtual Reality Teleoperation , 2017, ICRA.

[26]  Cynthia Breazeal,et al.  Designing sociable robots , 2002 .

[27]  Yaser Sheikh,et al.  Hand Keypoint Detection in Single Images Using Multiview Bootstrapping , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[28]  Chyi-Yeu Lin,et al.  An expressional simplified mechanism in anthropomorphic face robot design , 2016, Robotica.

[29]  Andrew Zisserman,et al.  X2Face: A network for controlling face generation by using images, audio, and pose codes , 2018, ECCV.

[30]  Cynthia Breazeal,et al.  Regulation and Entrainment in Human—Robot Interaction , 2000, Int. J. Robotics Res..

[31]  Cynthia Breazeal,et al.  Social Robots that Interact with People , 2008, Springer Handbook of Robotics.

[32]  Jun-Ho Oh,et al.  Design of Android type Humanoid Robot Albert HUBO , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[33]  Minoru Asada,et al.  Realistic child robot “Affetto” for understanding the caregiver-child attachment relationship that guides the child development , 2011, 2011 IEEE International Conference on Development and Learning (ICDL).

[34]  Jie Tan,et al.  Learning Agile Robotic Locomotion Skills by Imitating Animals , 2020, RSS 2020.

[35]  Nicu Sebe,et al.  First Order Motion Model for Image Animation , 2020, NeurIPS.

[36]  Goldie Nejat,et al.  How Robots Influence Humans: A Survey of Nonverbal Communication in Social Human–Robot Interaction , 2019, International Journal of Social Robotics.

[37]  Cynthia Breazeal,et al.  Toward Teaching a Robot "Infant" using Emotive Communication Acts , 1999 .

[38]  Min Wu,et al.  A facial expression emotion recognition based human-robot interaction system , 2017, IEEE/CAA Journal of Automatica Sinica.

[39]  Jan Kautz,et al.  Video-to-Video Synthesis , 2018, NeurIPS.

[40]  MengChu Zhou,et al.  Image Ratio Features for Facial Expression Recognition Application , 2010, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[41]  Jan Kautz,et al.  Few-shot Video-to-Video Synthesis , 2019, NeurIPS.

[42]  Martin A. Giese,et al.  Perceptual Robotics , 2008, Springer Handbook of Robotics, 2nd Ed..

[43]  Michael A. Goodrich,et al.  Human-Robot Interaction: A Survey , 2008, Found. Trends Hum. Comput. Interact..

[44]  Tamim Asfour,et al.  Imitation Learning of Dual-Arm Manipulation Tasks in Humanoid Robots , 2006, 2006 6th IEEE-RAS International Conference on Humanoid Robots.

[45]  Dong-Wook Lee,et al.  Optimizing Android Facial Expressions Using Genetic Algorithms , 2019, Applied Sciences.

[46]  Minoru Asada,et al.  Design of 22-DOF pneumatically actuated upper body for child android ‘Affetto’ , 2015, Adv. Robotics.

[47]  T. Chartrand,et al.  Where is the love? The social aspects of mimicry , 2009, Philosophical Transactions of the Royal Society B: Biological Sciences.

[48]  Hod Lipson,et al.  Facially expressive humanoid robotic face , 2020 .

[49]  Lior Wolf,et al.  Vid2Game: Controllable Characters Extracted from Real-World Videos , 2019, ICLR.

[50]  Kun Zhou,et al.  Displaced dynamic expression regression for real-time facial tracking and animation , 2014, ACM Trans. Graph..

[51]  Yaser Sheikh,et al.  VR facial animation via multiview image translation , 2019, ACM Trans. Graph..

[52]  Maya Cakmak,et al.  Characterizing the Design Space of Rendered Robot Faces , 2018, 2018 13th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[53]  Yaser Sheikh,et al.  OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[54]  Justus Thies,et al.  Real-time expression transfer for facial reenactment , 2015, ACM Trans. Graph..

[55]  Brian Scassellati,et al.  Infant-like Social Interactions between a Robot and a Human Caregiver , 2000, Adapt. Behav..

[56]  R. Brooks,et al.  The cog project: building a humanoid robot , 1999 .

[57]  Cynthia Breazeal,et al.  Emotion and sociable humanoid robots , 2003, Int. J. Hum. Comput. Stud..

[58]  A. Meltzoff,et al.  Imitation in Newborn Infants: Exploring the Range of Gestures Imitated and the Underlying Mechanisms. , 1989, Developmental psychology.

[59]  B. Scassellati Imitation and mechanisms of joint attention: a developmental structure for building social skills on a humanoid robot , 1999 .

[60]  David Lee,et al.  The art of designing robot faces: dimensions for human-robot interaction , 2006, HRI '06.

[61]  Haifeng Hu,et al.  Local robust sparse representation for face recognition with single sample per person , 2018, IEEE/CAA Journal of Automatica Sinica.

[62]  Alexei A. Efros,et al.  Everybody Dance Now , 2018, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[63]  Siddhartha S. Srinivasa,et al.  Imitation learning for locomotion and manipulation , 2007, 2007 7th IEEE-RAS International Conference on Humanoid Robots.

[64]  Thomas Brox,et al.  FlowNet: Learning Optical Flow with Convolutional Networks , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[65]  Maja Pantic,et al.  Web-based database for facial expression analysis , 2005, 2005 IEEE International Conference on Multimedia and Expo.

[66]  Marcelo H. Ang,et al.  A Survey on Perception Methods for Human–Robot Interaction in Social Robots , 2013, International Journal of Social Robotics.

[67]  Chen Fang,et al.  Dance Dance Generation: Motion Transfer for Internet Videos , 2019, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW).

[68]  Fuji Ren,et al.  Emotion classification using a CNN_LSTM-based model for smooth emotional synchronization of the humanoid robot REN-XIN , 2019, PloS one.

[69]  Judy Kay,et al.  Interactive Realistic Digital Avatars - Revisiting the Uncanny Valley , 2017, HICSS.

[70]  Illah R. Nourbakhsh,et al.  A survey of socially interactive robots , 2003, Robotics Auton. Syst..