Position/force control of robot manipulators using reinforcement learning

The position/force control of the robot needs the parameters of the impedance model and generates the desired position from the contact force in the environment. When the environment is unknown, learning algorithms are needed to estimate both the desired force and the parameters of the impedance model.,In this paper, the authors use reinforcement learning to learn only the desired force, then they use proportional-integral-derivative admittance control to generate the desired position. The results of the experiment are presented to verify their approach.,The position error is minimized without knowing the environment or the impedance parameters. Another advantage of this simplified position/force control is that the transformation of the Cartesian space to the joint space by inverse kinematics is avoided by the feedback control mechanism. The stability of the closed-loop system is proven.,The position error is minimized without knowing the environment or the impedance parameters. The stability of the closed-loop system is proven.

[1]  Shuzhi Sam Ge,et al.  Optimal Critic Learning for Robot Control in Time-Varying Environments , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[2]  Neville Hogan,et al.  Impedance Control: An Approach to Manipulation: Part I—Theory , 1985 .

[3]  John T. Wen,et al.  Stability analysis of position and force control for robot arms , 1991 .

[4]  John T. Wen,et al.  Integral force control with robustness enhancement , 1994 .

[5]  Nikos A. Aspragathos,et al.  Online Stability in Human-Robot Cooperation with Admittance Control , 2016, IEEE Transactions on Haptics.

[6]  Paolo Rocco,et al.  Impedance control for elastic joints industrial manipulators , 2004, IEEE Transactions on Robotics and Automation.

[7]  Michael I. Jordan,et al.  MASSACHUSETTS INSTITUTE OF TECHNOLOGY ARTIFICIAL INTELLIGENCE LABORATORY and CENTER FOR BIOLOGICAL AND COMPUTATIONAL LEARNING DEPARTMENT OF BRAIN AND COGNITIVE SCIENCES , 1996 .

[8]  Muhammad Tufail,et al.  Impedance control schemes for bilateral teleoperation , 2014, 2014 9th International Conference on Computer Science & Education.

[9]  Robert G. Bonitz,et al.  Internal force-based impedance control for cooperating manipulators , 1996, IEEE Trans. Robotics Autom..

[10]  Robert Babuska,et al.  Efficient Model Learning Methods for Actor–Critic Control , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[11]  Robert Babuska,et al.  Model learning actor-critic algorithms: Performance evaluation in a motion control task , 2012, 2012 IEEE 51st IEEE Conference on Decision and Control (CDC).

[12]  M. Buss,et al.  Techniques for environment parameter estimation during telemanipulation , 2008, 2008 2nd IEEE RAS & EMBS International Conference on Biomedical Robotics and Biomechatronics.

[13]  Rolf Johansson,et al.  Quadratic optimization of impedance control , 1994, Proceedings of the 1994 IEEE International Conference on Robotics and Automation.

[14]  Bruno Siciliano,et al.  Variable Impedance Control of Redundant Manipulators for Intuitive Human–Robot Physical Interaction , 2015, IEEE Transactions on Robotics.

[15]  Bruno Siciliano,et al.  A survey of robot interaction control schemes with experimental comparison , 1999 .

[16]  Keyvan Hashtrudi-Zaad,et al.  Optimization-based Robot Compliance Control: Geometric and Linear Quadratic Approaches , 2005, Int. J. Robotics Res..

[17]  Robert Babuska,et al.  Actor-Critic Control with Reference Model Learning , 2011, IFAC Proceedings Volumes.

[18]  Dan O. Popa,et al.  An analysis of some fundamental problems in adaptive control of force and impedance behavior: theory and experiments , 1995, IEEE Trans. Robotics Autom..

[19]  Max Q.-H. Meng,et al.  Impedance control with adaptation for robotic manipulations , 1991, IEEE Trans. Robotics Autom..