Learning haptic affordances from demonstration and human-guided exploration

In this paper, we present a system for learning haptic affordance models of complex manipulation skills. The goal of a haptic affordance model is to better complete tasks by characterizing what a particular object-action pair feels like. We use learning from demonstration to provide the robot with an example of a successful interaction with a given object. We then use environmental scaffolding to collect grounded examples (successes and unsuccessful "near misses") of the haptic data for the object-action pair, using a force/torque (F/T) sensor mounted at the wrist. From this we build one success Hidden Markov Model (HMM) and one "near-miss" HMM for each object-action pair. We evaluate this approach with five different actions over seven different objects to learn two specific affordances (open-able and scoop-able). We show that by building a library of object-action pairs for each affordance, we can successfully monitor a trajectory of haptic data to determine if the robot finds an affordance. This is supported through cross-validation on the object-action pairs, with four of the seven object-action pairs achieving a perfect F1 score, and with leave-one-object-out testing, where the learned object-actions models correctly identify the specific affordance with an average accuracy of 67% for scoop-able and 53% for open-able.

[1]  Chu Vivian,et al.  Learning object affordances by leveraging the combination of human-guidance and self-exploration , 2016 .

[2]  Ashutosh Saxena,et al.  Learning haptic representation for manipulating deformable food objects , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[3]  Alexander Stoytchev,et al.  Learning to slide a magnetic card through a card reader , 2012, 2012 IEEE International Conference on Robotics and Automation.

[4]  Darwin G. Caldwell,et al.  Imitation Learning of Positional and Force Skills Demonstrated via Kinesthetic Teaching and Haptic Input , 2011, Adv. Robotics.

[5]  M. Mascolo Change processes in development: The concept of coactive scaffolding , 2005 .

[6]  Aude Billard,et al.  Robot Learning from Failed Demonstrations , 2012, Int. J. Soc. Robotics.

[7]  Alexander Stoytchev,et al.  Behavior-Grounded Representation of Tool Affordances , 2005, Proceedings of the 2005 IEEE International Conference on Robotics and Automation.

[8]  Stefan Schaal,et al.  Learning and generalization of motor skills by learning from demonstration , 2009, 2009 IEEE International Conference on Robotics and Automation.

[9]  Wolfram Burgard,et al.  Operating articulated objects based on experience , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[10]  James M. Rehg,et al.  Rapid categorization of object properties from incidental contact with a tactile sensing robot arm , 2013, 2013 13th IEEE-RAS International Conference on Humanoid Robots (Humanoids).

[11]  Carme Torras,et al.  A robot learning from demonstration framework to perform force-based manipulation tasks , 2013, Intelligent Service Robotics.

[12]  Maya Cakmak,et al.  Using learned affordances for robotic behavior development , 2008, 2008 IEEE International Conference on Robotics and Automation.

[13]  Peter Stone,et al.  Transfer Learning for Reinforcement Learning Domains: A Survey , 2009, J. Mach. Learn. Res..

[14]  James M. Rehg,et al.  Decoupling behavior, perception, and control for autonomous learning of affordances , 2013, 2013 IEEE International Conference on Robotics and Automation.

[15]  Pierre-Yves Oudeyer,et al.  Intrinsic Motivation Systems for Autonomous Mental Development , 2007, IEEE Transactions on Evolutionary Computation.

[16]  Gerald E. Loeb,et al.  Bayesian Exploration for Intelligent Identification of Textures , 2012, Front. Neurorobot..

[17]  Maya Cakmak,et al.  To Afford or Not to Afford: A New Formalization of Affordances Toward Affordance-Based Robot Control , 2007, Adapt. Behav..

[18]  Stefan Schaal,et al.  Skill learning and task outcome prediction for manipulation , 2011, 2011 IEEE International Conference on Robotics and Automation.

[19]  Manuel Lopes,et al.  Learning Object Affordances: From Sensory--Motor Coordination to Imitation , 2008, IEEE Transactions on Robotics.

[20]  Connor Schenck,et al.  Interactive object recognition using proprioceptive and auditory feedback , 2011, Int. J. Robotics Res..

[21]  Lorenzo Natale,et al.  Tapping into Touch , 2005 .

[22]  Trevor Darrell,et al.  Using robotic exploratory procedures to learn the meaning of haptic adjectives , 2013, 2013 IEEE International Conference on Robotics and Automation.

[23]  Maya Cakmak,et al.  Keyframe-based Learning from Demonstration , 2012, Int. J. Soc. Robotics.

[24]  Stefan Schaal,et al.  Towards Associative Skill Memories , 2012, 2012 12th IEEE-RAS International Conference on Humanoid Robots (Humanoids 2012).

[25]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[26]  Giulio Sandini,et al.  Learning about objects through action - initial steps towards artificial cognition , 2003, 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422).

[27]  James M. Rehg,et al.  Inferring Object Properties from Incidental Contact with a Tactile Sensing Forearm , 2014, ArXiv.

[28]  Maya Cakmak,et al.  Learning about objects with human teachers , 2009, 2009 4th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[29]  Stefan Schaal,et al.  Online movement adaptation based on previous sensor experiences , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[30]  L. Rabiner,et al.  An introduction to hidden Markov models , 1986, IEEE ASSP Magazine.

[31]  Aude Billard,et al.  Donut as I do: Learning from failed demonstrations , 2011, 2011 IEEE International Conference on Robotics and Automation.