A Bayesian hierarchy for robust gaze estimation in human-robot interaction

In this text, we present a probabilistic solution for robust gaze estimation in the context of human–robot interaction. Gaze estimation, in the sense of continuously assessing gaze direction of an interlocutor so as to determine his/her focus of visual attention, is important in several important computer vision applications, such as the development of non-intrusive gaze-tracking equipment for psychophysical experiments in neuroscience, specialised telecommunication devices, video surveillance, human–computer interfaces (HCI) and artificial cognitive systems for human–robot interaction (HRI), our application of interest. We have developed a robust solution based on a probabilistic approach that inherently deals with the uncertainty of sensor models, but also and in particular with uncertainty arising from distance, incomplete data and scene dynamics. This solution comprises a hierarchical formulation in the form of a mixture model that loosely follows how geometrical cues provided by facial features are believed to be used by the human perceptual system for gaze estimation. A quantitative analysis of the proposed framework's performance was undertaken through a thorough set of experimental sessions. Results show that the framework performs according to the difficult requirements of HRI applications, namely by exhibiting correctness, robustness and adaptiveness.

[1]  E. Hall,et al.  The Hidden Dimension , 1970 .

[2]  S. Langton,et al.  The influence of head contour and nose angle on the perception of eye-gaze direction , 2004, Perception & psychophysics.

[3]  Steven K. Feiner,et al.  Gaze locking: passive eye contact detection for human-object interaction , 2013, UIST.

[4]  Henrik I. Christensen,et al.  Evaluation of Passing Distance for Social Robots , 2006, ROMAN 2006 - The 15th IEEE International Symposium on Robot and Human Interactive Communication.

[5]  Kamel Mekhnacha,et al.  Bayesian Programming , 2013 .

[6]  Nikos Fakotakis,et al.  An unconstrained method for lip detection in color images , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[7]  Jorge Dias,et al.  Attentional Mechanisms for Socially Interactive Robots–A Survey , 2014, IEEE Transactions on Autonomous Mental Development.

[8]  L. Whitaker Anthropometry of the Head and Face in Medicine. , 1983 .

[9]  Jorge Dias,et al.  A Bayesian framework for active artificial perception , 2013, IEEE Transactions on Cybernetics.

[10]  Olivier White,et al.  Computation of gaze orientation under unrestrained head movements , 2007, Journal of Neuroscience Methods.

[11]  Jean-Marc Odobez,et al.  Geometric Generative Gaze Estimation (G3E) for Remote RGB-D Cameras , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[12]  Niklas Bergström,et al.  Modeling of natural human-robot encounters , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[13]  Sander Oude Elberink,et al.  Accuracy and Resolution of Kinect Depth Data for Indoor Mapping Applications , 2012, Sensors.

[14]  Mayu Nishimura,et al.  What are you looking at? Acuity for triadic eye gaze. , 2004, The Journal of general psychology.

[15]  Paulo Menezes,et al.  An Interactive System for People Suffering from Cerebral Palsy , 2013 .

[16]  William Hyde Wollaston,et al.  XIII. On the apparent direction of eyes in a portrait , 1824, Philosophical Transactions of the Royal Society of London.

[17]  Nicu Sebe,et al.  Combining Head Pose and Eye Location Information for Gaze Estimation , 2012, IEEE Transactions on Image Processing.

[18]  Yoichi Sato,et al.  Appearance-Based Gaze Estimation Using Visual Saliency , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[19]  Matthew W. Hoffman,et al.  A probabilistic model of gaze imitation and shared attention , 2006, Neural Networks.

[20]  Jorge Dias,et al.  Designing an artificial attention system for social robots , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[21]  Pablo Lanillos,et al.  Fast Exact Bayesian Inference for High-Dimensional Models , 2015 .

[22]  Albert Ali Salah,et al.  Joint Attention by Gaze Interpolation and Saliency , 2013, IEEE Transactions on Cybernetics.

[23]  P. Thier,et al.  How precise is gaze following in humans? , 2008, Vision Research.

[24]  Pablo Lanillos,et al.  Gaze Tracing in a Bounded Log-Spherical Space for Artificial Attention Systems , 2015, ROBOT.

[25]  Robin R. Murphy,et al.  Evaluation of Proxemic Scaling Functions for Social Robotics , 2014, IEEE Transactions on Human-Machine Systems.

[26]  Takayuki Kanda,et al.  Spatial Formation Model for Initiating Conversation , 2011, Robotics: Science and Systems.

[27]  Dejan Todorović,et al.  Geometrical basis of perception of gaze direction , 2006, Vision Research.

[28]  Christopher R Forrest,et al.  International Anthropometric Study of Facial Morphology in Various Ethnic Groups/Races , 2005, The Journal of craniofacial surgery.

[29]  Jean-Marc Odobez,et al.  Gaze estimation from multimodal Kinect data , 2012, 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[30]  Dotan Knaan,et al.  Single image face orientation and gaze detection , 2009, Machine Vision and Applications.

[31]  Jorge Dias,et al.  Probabilistic Approaches to Robotic Perception , 2014, Springer Tracts in Advanced Robotics.

[32]  Radu Horaud,et al.  Simultaneous estimation of gaze direction and visual focus of attention for multi-person-to-robot interaction , 2016, 2016 IEEE International Conference on Multimedia and Expo (ICME).

[33]  Montse Pardàs,et al.  Head Orientation Estimation Using Particle Filtering in Multiview Scenarios , 2007, CLEAR.

[34]  David Portugal,et al.  Distributed multi-robot patrol: A scalable and fault-tolerant framework , 2013, Robotics Auton. Syst..

[35]  Takahiro Okabe,et al.  Inferring human gaze from appearance via adaptive linear regression , 2011, 2011 International Conference on Computer Vision.

[36]  Takayuki Kanda,et al.  How to approach humans?-strategies for social robots to initiate interaction , 2009, 2009 4th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[37]  Gigel Macesanu,et al.  A Time-Delay Control Approach for a Stereo Vision Based Human-Machine Interaction System , 2014, J. Intell. Robotic Syst..

[38]  Mohan M. Trivedi,et al.  Head Pose Estimation in Computer Vision: A Survey , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[39]  Takeo Kanade,et al.  Pose Robust Face Tracking by Combining Active Appearance Models and Cylinder Head Models , 2007, International Journal of Computer Vision.

[40]  L. Farkas,et al.  Comparison of anthropometric and cephalometric measurements of the adult face. , 1999, The Journal of craniofacial surgery.

[41]  Ioannis Pitas,et al.  Facial feature extraction and pose determination , 2000, Pattern Recognit..

[42]  Kostas Karpouzis,et al.  Visual Focus of Attention in Non-calibrated Environments using Gaze Estimation , 2014, International Journal of Computer Vision.