Learning Compliant Manipulation Tasks from Force Demonstrations

Robots work around human beings not only on manufacturing production lines, but also in work scenes such as kitchen, operating table and other unstructured environments. Preprogramming stiff position control is not enough. To exploit how to perform compliant manipulations around human beings without having to preprogram them is meaningful. In this paper, we aimed at endowing robot with the ability of understanding and learning compliant manipulation from demonstrations. In demonstration phase, the interaction forces and the end effector trajectory are recorded through a simple teaching interface. Furthermore, a learning framework is proposed to exploit the interaction force to estimate time-varying stiffness and damping profiles of the impedance controller in x, y and z directions. We validate the proposed approach with two experiments on a 4-DoF Barrett WAM, where the robot is able to learn time-varying compliance levels in three directions from demonstrations.

[1]  Evangelos Theodorou,et al.  Tendon-Driven Variable Impedance Control Using Reinforcement Learning , 2012, Robotics: Science and Systems.

[2]  Aude Billard,et al.  Online learning of varying stiffness through physical human-robot interaction , 2012, 2012 IEEE International Conference on Robotics and Automation.

[3]  Chao Xu,et al.  Fast and Stable Learning of Dynamical Systems Based on Extreme Learning Machine , 2019, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[4]  Ville Kyrki,et al.  Learning compliant assembly motions from demonstration , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[5]  Xinyu Wu,et al.  Neural learning of stable dynamical systems based on extreme learning machine , 2015, 2015 IEEE International Conference on Information and Automation.

[6]  Sethu Vijayakumar,et al.  Optimal variable stiffness control: formulation and application to explosive movement tasks , 2012, Auton. Robots.

[7]  Sandra Hirche,et al.  Risk-sensitive interaction control in uncertain manipulation tasks , 2013, 2013 IEEE International Conference on Robotics and Automation.

[8]  Stefan Schaal,et al.  Learning variable impedance control , 2011, Int. J. Robotics Res..

[9]  Darwin G. Caldwell,et al.  Learning-based control strategy for safe human-robot interaction exploiting task and robot redundancies , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[10]  Bruno Siciliano,et al.  Task-Space Control of Robot Manipulators With Null-Space Compliance , 2014, IEEE Transactions on Robotics.

[11]  Darwin G. Caldwell,et al.  A Method for Derivation of Robot Task-Frame Control Authority from Repeated Sensory Observations , 2017, IEEE Robotics and Automation Letters.

[12]  Jun Nakanishi,et al.  Learning Attractor Landscapes for Learning Motor Primitives , 2002, NIPS.

[13]  Darwin G. Caldwell,et al.  Robot motor skill coordination with EM-based Reinforcement Learning , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[14]  Nicholas Roy,et al.  Tendon-Driven Variable Impedance Control Using Reinforcement Learning , 2013 .

[15]  Aude Billard,et al.  Learning control Lyapunov function to ensure stability of dynamical system-based robot reaching motions , 2014, Robotics Auton. Syst..

[16]  Aude Billard,et al.  Learning Compliant Manipulation through Kinesthetic and Tactile Human-Robot Interaction , 2014, IEEE Transactions on Haptics.

[17]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[18]  Aude Billard,et al.  Learning Stable Nonlinear Dynamical Systems With Gaussian Mixture Models , 2011, IEEE Transactions on Robotics.

[19]  Mitsuo Kawato,et al.  Equilibrium-Point Control Hypothesis Examined by Measured Arm Stiffness During Multijoint Movement , 1996, Science.

[20]  Aude Billard,et al.  Learning from Humans , 2016, Springer Handbook of Robotics, 2nd Ed..