Approaches for Learning Human-like Motor Skills which Require Variable Stiffness During Execution

Humans employ varying stiffness in everyday life for almost all human motor skills, using both passive and active compliance. Robots have only recently acquired variable passive stiffness actuators and they are not yet mature. Active compliance controllers have existed for a longer time, but the problem of automatic determination of the necessary compliance to achieve a task has not been thoroughly studied. Teaching humanoid robots to apply variable stiffness to the skills they acquire is vital in order to achieve human-like naturalness of the execution. Also, using adaptive compliance can help to increase the energy efficiency. This paper compares two different approaches that allow robots to learn human-like skills which require varying stiffness during execution. The advantages and disadvantages of each approach is discussed and demonstrated with various experiments on an activelycompliant Barrett WAM robot. keywords: variable stiffness, motor skills, imitation learning, reinforcement learning, programming by demonstration, kinesthetic teaching, pHRI

[1]  Aude Billard,et al.  BM: An iterative algorithm to learn stable non-linear dynamical systems with Gaussian mixture models , 2010, 2010 IEEE International Conference on Robotics and Automation.

[2]  Eric L. Sauser,et al.  An Approach Based on Hidden Markov Model and Gaussian Mixture Regression , 2010 .

[3]  Pradeep K. Khosla,et al.  Exploiting redundancy to reduce impact force , 1991, Proceedings IROS '91:IEEE/RSJ International Workshop on Intelligent Robots and Systems '91.

[4]  Jan Peters,et al.  Learning motor primitives for robotics , 2009, 2009 IEEE International Conference on Robotics and Automation.

[5]  Henk Nijmeijer,et al.  Robot Programming by Demonstration , 2010, SIMPAR.

[6]  Darwin G. Caldwell,et al.  Robot motor skill coordination with EM-based Reinforcement Learning , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[7]  Pradeep K. Khosla,et al.  Exploiting redundancy to reduce impact force , 1994, J. Intell. Robotic Syst..

[8]  Brett Browning,et al.  A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..

[9]  Darwin G. Caldwell,et al.  Learning and Reproduction of Gestures by Imitation , 2010, IEEE Robotics & Automation Magazine.

[10]  Alessandro De Luca,et al.  Exploiting robot redundancy in collision detection and reaction , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[11]  B. Schölkopf,et al.  Reinforcement Learning for Motor Primitives , 2009 .

[12]  Stefan Schaal,et al.  Robot Programming by Demonstration , 2009, Springer Handbook of Robotics.

[13]  Sethu Vijayakumar,et al.  Methods for Learning Control Policies from Variable-Constraint Demonstrations , 2010, From Motor Learning to Interaction Learning in Robots.

[14]  Stefan Schaal,et al.  Dynamics systems vs. optimal control--a unifying view. , 2007, Progress in brain research.

[15]  Jun Nakanishi,et al.  Trajectory formation for imitation with nonlinear dynamical systems , 2001, Proceedings 2001 IEEE/RSJ International Conference on Intelligent Robots and Systems. Expanding the Societal Role of Robotics in the the Next Millennium (Cat. No.01CH37180).

[16]  Darwin G. Caldwell,et al.  Handling of multiple constraints and motion alternatives in a robot programming by demonstration framework , 2009, 2009 9th IEEE-RAS International Conference on Humanoid Robots.

[17]  Stefan Schaal,et al.  Biologically-inspired dynamical systems for movement generation: Automatic real-time goal adaptation and obstacle avoidance , 2009, 2009 IEEE International Conference on Robotics and Automation.