Automatic demonstration and feature selection for robot learning

Robot learning frameworks, such as Programming by Demonstration, are based on learning tasks from sets of user demonstrations. These frameworks, in their naive implementation, assume that all the data from the user demonstrations has been correctly sensed and can be relevant to the task. Analogous to feature selection, which is the process of selecting a subset of relevant features for use in model construction, this paper presents a demonstration selection process, which is additionally applied for feature selection for further data filtering. The demonstration and feature selection process presented is called Dissimilarity Mapping Filtering (DMF). DMF involves three steps: obtaining a measurement of dissimilarity (e.g. Dynamic Time Warping, etc.), reducing dimensions through a mapping algorithm (e.g. sum of dissimilarities, Multidimensional Scaling, etc.) and a filtering method (z-score based, DBSCAN, etc.). As a demonstration selector, DMF discards outlying demonstrations in terms of all the features considered simultaneously. As a feature selector, DMF discards features that present high inconsistency among demonstrations. We apply DMF to our Continuous Goal-Directed Actions (CGDA) robot learning framework presented in previous works.

[1]  Sonia Garcia-Salicetti,et al.  Dynamic Time Warping (DTW) , 2009, Encyclopedia of Biometrics.

[2]  Jochen J. Steil,et al.  Task-level imitation learning using variance-based movement optimization , 2009, 2009 IEEE International Conference on Robotics and Automation.

[3]  Carlos Balaguer,et al.  On using guided motor primitives to execute Continuous Goal-Directed Actions , 2014, The 23rd IEEE International Symposium on Robot and Human Interactive Communication.

[4]  Carlos Balaguer,et al.  Action effect generalization, recognition and execution through Continuous Goal-Directed Actions , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[5]  Carlos Balaguer,et al.  Humanoid robot imitation through continuous goal-directed actions: an evolutionary approach , 2015, Adv. Robotics.

[6]  Manuela M. Veloso,et al.  Multi-thresholded approach to demonstration selection for interactive robot learning , 2008, 2008 3rd ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[7]  Stefan Schaal,et al.  Robot Learning From Demonstration , 1997, ICML.

[8]  Sylvain Calino,et al.  Robot programming by demonstration : a probabilistic approach , 2009 .

[9]  Meinard Müller,et al.  Dynamic Time Warping , 2008 .

[11]  S. Schaal Dynamic Movement Primitives -A Framework for Motor Control in Humans and Humanoid Robotics , 2006 .

[12]  Carlos Balaguer,et al.  TEO: FULL-SIZE HUMANOID ROBOT DESIGN POWERED BY A FUEL CELL SYSTEM , 2012, Cybern. Syst..

[13]  Marc Toussaint,et al.  Task Space Retrieval Using Inverse Feedback Control , 2011, ICML.

[14]  Aude Billard,et al.  Goal-Directed Imitation in a Humanoid Robot , 2005, Proceedings of the 2005 IEEE International Conference on Robotics and Automation.

[15]  Jochen J. Steil,et al.  Automatic selection of task spaces for imitation learning , 2009, 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems.