Stochastic Genetic Algorithm-Assisted Fuzzy Q-Learning for Robotic Manipulators

This work proposes stochastic genetic algorithm-assisted Fuzzy Q-Learning-based robotic manipulator control. Specifically, the aim is to redefine the action choosing mechanism in Fuzzy Q-Learning for robotic manipulator control. Conventionally, a Fuzzy Q-Learning-based controller selects a deterministic action from available actions using fuzzy Q values. This deterministic Fuzzy Q-Learning is not an efficient approach, especially in dealing with highly coupled nonlinear systems such as robotic manipulators. Restricting the search for optimal action to the agent’s action set or a restricted set of Q values (deterministic) is a myopic idea. Herein, the proposal is to employ genetic algorithm as stochastic optimizer for action selection at each stage of Fuzzy Q-Learning-based controller. This turns out to be a highly effective way for robotic manipulator control rather than choosing an algebraic minimal action. As case studies, present work implements the proposed approach on two manipulators: (a) two-link arm manipulator and (b) selective compliance assembly robotic arm. Scheme is compared with baseline Fuzzy Q-Learning controller, Lyapunov Markov game-based controller and Linguistic Lyapunov Reinforcement Learning controller. Simulation results show that our stochastic genetic algorithm-assisted Fuzzy Q-Learning controller outperforms the above-mentioned controllers in terms of tracking errors along with lower torque requirements.

[1]  Andreas Krause,et al.  Safe Model-based Reinforcement Learning with Stability Guarantees , 2017, NIPS.

[2]  Rajneesh Sharma,et al.  Genetic Algorithm-Optimized Fuzzy Lyapunov Reinforcement Learning for Nonlinear Systems , 2020, Arabian Journal for Science and Engineering.

[3]  Wei He,et al.  Admittance-Based Controller Design for Physical Human–Robot Interaction in the Constrained Task Space , 2020, IEEE Transactions on Automation Science and Engineering.

[4]  Wei He,et al.  Adaptive Fuzzy Full-State and Output-Feedback Control for Uncertain Robots With Output Constraint , 2021, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[5]  Bin Fang,et al.  NAO robot obstacle avoidance based on fuzzy Q-learning , 2019, Ind. Robot.

[6]  B. A. Qureshi,et al.  A Comprehensive Review of Saline Water Correlations and Data: Part II—Thermophysical Properties , 2020, Arabian Journal for Science and Engineering.

[7]  Radu-Emil Precup,et al.  Model-Free control performance improvement using virtual reference feedback tuning and reinforcement Q-learning , 2017, Int. J. Syst. Sci..

[8]  Rajneesh Sharma,et al.  Fuzzy Lyapunov Reinforcement Learning for Non Linear Systems. , 2017, ISA transactions.

[9]  Hassan Bevrani,et al.  Load–frequency control : a GA-based multi-agent reinforcement learning , 2010 .

[10]  Shuai Li,et al.  Robot manipulator control using neural networks: A survey , 2018, Neurocomputing.

[11]  Changyin Sun,et al.  Neural Network Control of a Two-Link Flexible Robotic Manipulator Using Assumed Mode Method , 2019, IEEE Transactions on Industrial Informatics.

[12]  Javier González,et al.  Towards a common implementation of reinforcement learning for multiple robotic tasks , 2017, Expert Syst. Appl..

[13]  Francesco Braghin,et al.  Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration , 2020, Journal of Intelligent & Robotic Systems.

[14]  Tingwen Huang,et al.  Model-Free Optimal Tracking Control via Critic-Only Q-Learning , 2016, IEEE Transactions on Neural Networks and Learning Systems.

[15]  Rajneesh Sharma,et al.  Lyapunov fuzzy Markov game controller for two link robotic manipulator , 2018, J. Intell. Fuzzy Syst..

[16]  Rajneesh Sharma,et al.  Linguistic Lyapunov reinforcement learning control for robotic manipulators , 2018, Neurocomputing.