Bootstrapping humanoid robot skills by extracting semantic representations of human-like activities from virtual reality

Advancements in Virtual Reality have enabled well-defined and consistent virtual environments that can capture complex scenarios, such as human everyday activities. Additionally, virtual simulators (such as SIGVerse) are designed to be user-friendly mechanisms between virtual robots/agents and real users allowing a better interaction. We envision such rich scenarios can be used to train robots to learn new behaviors specially in human everyday activities where a diverse variability can be found. In this paper, we present a multi-level framework that is capable to use different input sources such as cameras and virtual environments to understand and execute the demonstrated activities. Our presented framework first obtains the semantic models of human activities from cameras, which are later tested using the SIGVerse virtual simulator to show new complex activities (such as, cleaning the table) using a virtual robot. Our introduced framework is integrated on a real robot, i.e. an iCub, which is capable to process the signals from the virtual environment to then understand the activities performed by the observed robot. This was realized through the use of previous knowledge and experiences that the robot has learned from observing humans activities. Our results show that our framework was able to extract the meaning of the observed motions with 80% accuracy of recognition by obtaining the objects relationships given the current context via semantic representations to extract high-level understanding of those complex activities even when they represent different behaviors.

[1]  Gordon Cheng,et al.  A full body human motion capture system using particle filtering and on-the-fly edge detection , 2004, 4th IEEE/RAS International Conference on Humanoid Robots, 2004..

[2]  Byoung-Tak Zhang,et al.  Enhancing human action recognition through spatio-temporal feature learning and semantic rules , 2013, 2013 13th IEEE-RAS International Conference on Humanoid Robots (Humanoids).

[3]  Gordon Cheng,et al.  Extracting Semantic Rules from Human Observations , 2013, ICRA 2013.

[4]  Katsushi Ikeuchi,et al.  Toward assembly plan from observation - Task recognition with planar, curved and mechanical contacts , 1993, Proceedings of 1993 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS '93).

[5]  Deva Ramanan,et al.  Detecting activities of daily living in first-person camera views , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[6]  Hema Swetha Koppula,et al.  Anticipating Human Activities Using Object Affordances for Reactive Robotic Response , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  Eren Erdal Aksoy,et al.  Learning the semantics of object–action relations by observation , 2011, Int. J. Robotics Res..

[8]  Gordon Cheng,et al.  Learning Similar Tasks From Observation and Practice , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[9]  J.K. Aggarwal,et al.  Human activity analysis , 2011, ACM Comput. Surv..

[10]  Yuichiro Yoshikawa,et al.  Simulator platform that enables social interaction simulation — SIGVerse: SocioIntelliGenesis simulator , 2010, 2010 IEEE/SICE International Symposium on System Integration.

[11]  Ronald Poppe,et al.  A survey on vision-based human action recognition , 2010, Image Vis. Comput..

[12]  Naoto Iwahashi,et al.  Unsupervised Segmentation of Human Motion Data Using a Sticky Hierarchical Dirichlet Process-Hidden Markov Model and Minimal Description Length-Based Chunking Method for Imitation Learning , 2011, Adv. Robotics.

[13]  Anthony G. Cohn,et al.  Learning Functional Object-Categories from a Relational Spatio-Temporal Representation , 2008, ECAI.

[14]  Rüdiger Dillmann,et al.  Teaching and learning of robot tasks via observation of human performance , 2004, Robotics Auton. Syst..

[15]  Aiko M. Hormann,et al.  Programs for Machine Learning. Part I , 1962, Inf. Control..

[16]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[17]  Danica Kragic,et al.  Learning Actions from Observations , 2010, IEEE Robotics & Automation Magazine.

[18]  Michael Ulbrich,et al.  Imitating human reaching motions using physically inspired optimization principles , 2011, 2011 11th IEEE-RAS International Conference on Humanoid Robots.

[19]  Gordon Cheng,et al.  Automatic segmentation and recognition of human activities from observation based on semantic reasoning , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[20]  Giulio Sandini,et al.  The iCub humanoid robot: an open platform for research in embodied cognition , 2008, PerMIS.

[21]  Henry A. Kautz,et al.  Inferring activities from interactions with objects , 2004, IEEE Pervasive Computing.

[22]  Yoshihiko Nakamura,et al.  Humanoid Robot's Autonomous Acquisition of Proto-Symbols through Motion Segmentation , 2006, 2006 6th IEEE-RAS International Conference on Humanoid Robots.

[23]  Moritz Tenorth,et al.  Towards Automated Models of Activities of Daily Life , 2009 .

[24]  Jeffrey Too Chuan Tan,et al.  Development of RoboCup@Home Simulation towards Long-term Large Scale HRI , 2013, RoboCup.

[25]  F. Wörgötter,et al.  Florentin Wörgötter action relations by observation − Learning the semantics of object , 2011 .

[26]  Tamim Asfour,et al.  Action sequence reproduction based on automatic segmentation and Object-Action Complexes , 2013, 2013 13th IEEE-RAS International Conference on Humanoid Robots (Humanoids).