Learning by demonstration for tool-handling task

When the robot comes to the home-like environment, its programming becomes very demanding. The concept of learning by demonstration is thus introduced, which intends to remove the load of detailed analysis and programming from the user. Following this concept, in this paper, we propose a novel approach for the robot to deduce the intention of the demonstrator from the trajectories during task execution. We focus on the tool-handling task, which is common in the home environment, but complicated for analysis. The proposed approach does not predefine motions or put constraints on motion speed, while allowing the event order to be altered and the presence of redundant operations during demonstration. We apply the concept of cross-validation to locate the portions of the trajectory that correspond to delicate and skillful maneuvering, and apply an algorithm based on dynamic programming previously developed to search for the most probable intention. In experiments, we apply the proposed approach for two different kinds of tasks, the pouring and coffee-making tasks, with the number of objects and their locations varied during demonstrations.

[1]  Rüdiger Dillmann,et al.  Incremental Learning of Tasks From User Demonstrations, Past Experiences, and Vocal Comments , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[2]  S. Ekvall,et al.  Integrating object and grasp recognition for dynamic scene interpretation , 2005, ICAR '05. Proceedings., 12th International Conference on Advanced Robotics, 2005..

[3]  Kuu-Young Young,et al.  Intention Learning From Human Demonstration , 2011, J. Inf. Sci. Eng..

[4]  Stefano Caselli,et al.  Toward programming of assembly tasks by demonstration in virtual environments , 2003, The 12th IEEE International Workshop on Robot and Human Interactive Communication, 2003. Proceedings. ROMAN 2003..

[5]  Minoru Asada,et al.  Automatic extraction of abstract actions from humanoid motion data , 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566).

[6]  Rüdiger Dillmann,et al.  Incremental Learning of Task Sequences with Information-Theoretic Metrics , 2006, EUROS.

[7]  Stefano Caselli,et al.  Leveraging on a virtual environment for robot programming by demonstration , 2004, Robotics Auton. Syst..

[8]  R. K. Shyamasundar,et al.  Introduction to algorithms , 1996 .

[9]  Stefan Schaal,et al.  Robot Programming by Demonstration , 2009, Springer Handbook of Robotics.

[10]  Hiroshi Kimura,et al.  Extraction of fine motion through multiple observations of human demonstration by DP matching and combined template matching , 2001, Proceedings 10th IEEE International Workshop on Robot and Human Interactive Communication. ROMAN 2001 (Cat. No.01TH8591).

[11]  Jodie A. Baird,et al.  Discerning intentions in dynamic human action , 2001, Trends in Cognitive Sciences.

[12]  Chrystopher L. Nehaniv,et al.  Sensory-Motor Primitives as a Basis for Imitation: Linking Perception to Action and Biology to Robotics , 2002 .

[13]  S. Chiba,et al.  Dynamic programming algorithm optimization for spoken word recognition , 1978 .

[14]  Tamim Asfour,et al.  Imitation Learning of Dual-Arm Manipulation Tasks in Humanoid Robots , 2006, 2006 6th IEEE-RAS International Conference on Humanoid Robots.

[15]  W. Marsden I and J , 2012 .

[16]  Aude Billard,et al.  On Learning, Representing, and Generalizing a Task in a Humanoid Robot , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[17]  Katsushi Ikeuchi,et al.  Extraction of essential interactions through multiple observations of human demonstrations , 2003, IEEE Trans. Ind. Electron..

[18]  K. Dautenhahn,et al.  The correspondence problem , 2002 .

[19]  E. John,et al.  Evoked-Potential Correlates of Stimulus Uncertainty , 1965, Science.

[20]  Kerstin Dautenhahn,et al.  Achieving Corresponding Effects on Multiple Robotic Platforms: Imitating in Context Using Different Effect Metrics , 2005 .

[21]  Masayuki Inaba,et al.  Learning by watching: extracting reusable task knowledge from visual observation of human performance , 1994, IEEE Trans. Robotics Autom..

[22]  W. Wong,et al.  On ψ-Learning , 2003 .