Social behavior recognition using body posture and head pose for human-robot interaction

Robots that interact with humans in everyday situations, need to be able to interpret the nonverbal social cues of their human interaction partners. We show that humans use body posture and head pose as social signals to initiate and terminate interaction when ordering drinks at a bar. For that, we record and analyze 108 interactions of humans interacting with a human bartender. Based on these findings, we train a Hidden Markov Model (HMM) using automatic body posture and head pose estimation. With this model, the bartender robot of the project JAMES can recognize typical social behaviors of human customers. Evaluation shows a recognition rate of 82.9 % for all implemented social behaviors and in particular a recognition rate of 91.2 % for bartender attention requests, which will allow the robot to interact with multiple humans in a robust and socially appropriate way.

[1]  Seong-Whan Lee,et al.  Gesture Spotting and Recognition for Human–Robot Interaction , 2007, IEEE Transactions on Robotics.

[2]  Ana Paiva,et al.  Affect recognition for interactive companions: challenges and design in real world scenarios , 2009, Journal on Multimodal User Interfaces.

[3]  Emanuel A. Schegloff Opening sequencing , 2002 .

[4]  Mark Aakhus,et al.  Perpetual Contact: Mobile Communication, Private Talk, Public Performance , 2002 .

[5]  喜多 壮太郎 Pointing : where language, culture, and cognition meet , 2013 .

[6]  Alois Knoll,et al.  Human workflow analysis using 3D occupancy grid hand tracking in a human-robot collaboration scenario , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[7]  Maja Pantic,et al.  Social signal processing: Survey of an emerging domain , 2009, Image Vis. Comput..

[8]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[9]  Alois Knoll,et al.  Modelling State of Interaction from Head Poses for Social Human-Robot Interaction , 2012, HRI 2012.

[10]  Peter Wittenburg,et al.  Annotation by Category: ELAN and ISO DCR , 2008, LREC.

[11]  H. H. Clark Pointing and placing. , 2003 .

[12]  Junji Yamato,et al.  Recognizing human action in time-sequential images using hidden Markov model , 1992, Proceedings 1992 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[13]  Matthias Scheutz,et al.  Toward Affective Cognitive Robots for Human-Robot Interaction , 2005, AAAI.

[14]  Haizhou Li,et al.  Experiences with a Barista Robot, FusionBot , 2009, FIRA RoboWorld Congress.

[15]  T. M. Ciolek,et al.  Environment and the Spatial Arrangement of Conversational Encounters , 1980 .

[16]  Christian Heath,et al.  Structures of Social Action: Talk and recipiency: sequential organization in speech and body movement , 1985 .

[17]  Martin Hägele,et al.  Robotic home assistant Care-O-bot® 3 - product vision and innovation platform , 2009, 2009 IEEE Workshop on Advanced Robotics and its Social Impacts.

[18]  Sven Behnke,et al.  Feature-based head pose estimation from images , 2007, 2007 7th IEEE-RAS International Conference on Humanoid Robots.

[19]  Paul A. Viola,et al.  Robust Real-Time Face Detection , 2001, International Journal of Computer Vision.