Reinforcement Learning Tracking Control for Robotic Manipulator With Kernel-Based Dynamic Model

Reinforcement learning (RL) is an efficient learning approach to solving control problems for a robot by interacting with the environment to acquire the optimal control policy. However, there are many challenges for RL to execute continuous control tasks. In this article, without the need to know and learn the dynamic model of a robotic manipulator, a kernel-based dynamic model for RL is proposed. In addition, a new tuple is formed through kernel function sampling to describe a robotic RL control problem. In this algorithm, a reward function is defined according to the features of tracking control in order to speed up the learning process, and then an RL tracking controller with a kernel-based transition dynamic model is proposed. Finally, a critic system is presented to evaluate the policy whether it is good or bad to the RL control tasks. The simulation results illustrate that the proposed method can fulfill the robotic tracking tasks effectively and achieve similar and even better tracking performance with much smaller inputs of force/torque compared with other learning algorithms, demonstrating the effectiveness and efficiency of the proposed RL algorithm.

[1]  Okyay Kaynak,et al.  Tracking Control of Robotic Manipulators With Uncertain Kinematics and Dynamics , 2016, IEEE Transactions on Industrial Electronics.

[2]  Yazhou Hu,et al.  A Reinforcement Learning Neural Network for Robotic Manipulator Control , 2018, Neural Computation.

[3]  André da Motta Salles Barreto,et al.  Practical Kernel-Based Reinforcement Learning , 2014, J. Mach. Learn. Res..

[4]  Frank L. Lewis,et al.  Adaptive H∞ Tracking Control of Nonlinear Systems Using Reinforcement Learning , 2018 .

[5]  Frank L. Lewis,et al.  Reinforcement Learning and Approximate Dynamic Programming for Feedback Control , 2012 .

[6]  Renquan Lu,et al.  Adaptive Neural Network Tracking Control for Robotic Manipulators With Dead Zone , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[7]  Quanmin Zhu,et al.  Adaptive synchronised tracking control for multiple robotic manipulators with uncertain kinematics and dynamics , 2016, Int. J. Syst. Sci..

[8]  Önder Tutsoy,et al.  Cpg Based RL Algorithm Learns to Control of a humanoid robot leg , 2015, Int. J. Robotics Autom..

[9]  P. Lions Optimal control of diffusion processes and hamilton–jacobi–bellman equations part 2 : viscosity solutions and uniqueness , 1983 .

[10]  Chi Zhang,et al.  Trajectory tracking control for rotary steerable systems using interval type-2 fuzzy logic and reinforcement learning , 2017, J. Frankl. Inst..

[11]  A. B. Rad,et al.  Robust fuzzy tracking control for robotic manipulators , 2007, Simul. Model. Pract. Theory.

[12]  Qiuye Sun,et al.  Nonlinear neuro-optimal tracking control via stable iterative Q-learning algorithm , 2015, Neurocomputing.

[13]  Derong Liu,et al.  Adaptive dynamic programming for robust neural control of unknown continuous-time non-linear systems , 2017 .

[14]  Sridhar Mahadevan,et al.  Average reward reinforcement learning: Foundations, algorithms, and empirical results , 2004, Machine Learning.

[15]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[16]  Miroslaw Galicki,et al.  Finite-time trajectory tracking control in a task space of robotic manipulators , 2016, Autom..

[17]  Sergey Levine,et al.  Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[18]  Huaguang Zhang,et al.  Online optimal tracking control of continuous-time linear systems with unknown dynamics by using adaptive dynamic programming , 2014, Int. J. Control.

[19]  Masato Sato,et al.  Robust Motion Control of an Oscillatory-Base Manipulator in a Global Coordinate System , 2015, IEEE Transactions on Industrial Electronics.

[20]  Mitsuo Kawato,et al.  Multiple Model-Based Reinforcement Learning , 2002, Neural Computation.

[21]  Sergey Levine,et al.  Temporal Difference Models: Model-Free Deep RL for Model-Based Control , 2018, ICLR.

[22]  Athanasios S. Polydoros,et al.  Survey of Model-Based Reinforcement Learning: Applications on Robotics , 2017, J. Intell. Robotic Syst..

[23]  Alexander J. Smola,et al.  Learning with kernels , 1998 .

[24]  Shen Yin,et al.  Exponential Tracking Control of Robotic Manipulators With Uncertain Dynamics and Kinematics , 2019, IEEE Transactions on Industrial Informatics.

[25]  Chao Chen,et al.  Adaptive Partial Reinforcement Learning Neural Network-Based Tracking Control for Wheeled Mobile Robotic Systems , 2020, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[26]  Akhilesh Swarup,et al.  Chattering Free Trajectory Tracking Control of a Robotic Manipulator Using High Order Sliding Mode , 2017 .

[27]  Abolfazl Mohebbi,et al.  Augmented Image-Based Visual Servoing of a Manipulator Using Acceleration Command , 2014, IEEE Transactions on Industrial Electronics.

[28]  Haibo He,et al.  Kernel-Based Approximate Dynamic Programming for Real-Time Online Learning Control: An Experimental Study , 2014, IEEE Transactions on Control Systems Technology.

[29]  Derong Liu,et al.  Adaptive Dynamic Programming for Optimal Tracking Control of Unknown Nonlinear Systems With Application to Coal Gasification , 2014, IEEE Transactions on Automation Science and Engineering.

[30]  Nicholas K. Jong,et al.  Kernel-Based Models for Reinforcement Learning , 2006 .