Human-robot Interaction Method Combining Human Pose Estimation and Motion Intention Recognition

Although human pose estimation technology based on RGB images is becoming more and more mature, most of the current mainstream methods rely on depth camera to obtain human joints information. These interaction frameworks are affected by the infrared detection distance so that they cannot well adapt to the interaction scene of different distance. Therefore, the purpose of this paper is to build a modular interactive framework based on RGB images, which aims to alleviate the problem of high dependence on depth camera and low adaptability to distance in the current human-robot interaction (HRI) framework based on human body by using advanced human pose estimation technology. To enhance the adaptability of the HRI framework to different distances, we adopt optical cameras instead of depth cameras as acquisition equipment. Firstly, the human joints information is extracted by a human pose estimation network. Then, a joints sequence filter is designed in the intermediate stage to reduce the influence of unreasonable skeletons on the interaction results. Finally, a human intention recognition model is built to recognize the human intention from reasonable joints information, and drive the robot to respond according to the predicted intention. The experimental results show that our interactive framework is more robust in the distance than the framework based on depth camera and is able to achieve effective interaction under different distances, illuminations, costumes, customers, and scenes.

[1]  Panos E. Trahanias,et al.  Gesture recognition based on arm tracking for human-robot interaction , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[2]  Osama Mazhar,et al.  A real-time human-robot interaction framework with robust background invariant hand gesture detection , 2019, Robotics Comput. Integr. Manuf..

[3]  Wei He,et al.  Bayesian Estimation of Human Impedance and Motion Intention for Human–Robot Collaboration , 2019, IEEE Transactions on Cybernetics.

[4]  Antonio Bicchi,et al.  Measuring intent in human-robot cooperative manipulation , 2009, 2009 IEEE International Workshop on Haptic Audio visual Environments and Games.

[5]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[6]  Dan Zhang,et al.  A Human-Robot Interaction for a Mecanum Wheeled Mobile Robot with Real-Time 3D Two-Hand Gesture Recognition , 2019, Journal of Physics: Conference Series.

[7]  Roland Siegwart,et al.  Kinect v2 for mobile robot navigation: Evaluation and modeling , 2015, 2015 International Conference on Advanced Robotics (ICAR).

[8]  Bernhard Schölkopf,et al.  Probabilistic movement modeling for intention inference in human–robot interaction , 2013, Int. J. Robotics Res..

[9]  Max Q.-H. Meng,et al.  Human robot cooperation based on human intention inference , 2014, 2014 IEEE International Conference on Robotics and Biomimetics (ROBIO 2014).

[10]  Ivan Markovic,et al.  Human Intention Recognition in Flexible Robotized Warehouses Based on Markov Decision Processes , 2017, ROBOT.

[11]  Trevor Darrell,et al.  Long-term recurrent convolutional networks for visual recognition and description , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Sergio Escalera,et al.  A real-time Human-Robot Interaction system based on gestures for assistive scenarios , 2016, Comput. Vis. Image Underst..

[13]  Xing Li,et al.  Human-robot interaction based on gesture and movement recognition , 2020, Signal Process. Image Commun..

[14]  Weihua Sheng,et al.  Wearable sensors based human intention recognition in smart assisted living systems , 2008, 2008 International Conference on Information and Automation.

[15]  Min Tan,et al.  Real-Time Human-Robot Interaction for a Service Robot Based on 3D Human Activity Recognition and Human-Mimicking Decision Mechanism , 2018, 2018 IEEE 8th Annual International Conference on CYBER Technology in Automation, Control, and Intelligent Systems (CYBER).

[16]  Jos Elfring,et al.  Learning intentions for improved human motion prediction , 2013, 2013 16th International Conference on Advanced Robotics (ICAR).

[17]  Fatik Baran Mandal Nonverbal Communication in Humans , 2014 .

[18]  Hema Swetha Koppula,et al.  Anticipating Human Activities Using Object Affordances for Reactive Robotic Response , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[19]  Robert X. Gao,et al.  Symbiotic human-robot collaborative assembly , 2019, CIRP Annals.

[20]  Yichen Wei,et al.  Towards 3D Human Pose Estimation in the Wild: A Weakly-Supervised Approach , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).