An iterative Q-learning based global consensus of discrete-time saturated multi-agent systems.

This paper addresses the consensus problem of discrete-time multiagent systems (DTMASs), which are subject to input saturation and lack of the information of agent dynamics. In the previous works, the DTMASs with input saturation can achieve semiglobal consensus by utilizing the low gain feedback (LGF) method, but computing the LGF matrices by solving the modified algebraic Riccati equation requires the knowledge of agent dynamics. In this paper, motivated by the reinforcement learning method, we propose a model-free Q-learning algorithm to obtain the LGF matrices for the DTMASs achieving global consensus. Firstly, we define a Q-learning function and deduce a Q-learning Bellman equation, whose solution can work out the LGF matrix. Then, we develop an iterative Q-learning algorithm to obtain the LGF matrix without the requirement of the knowledge about agent dynamics. Moreover, the DTMASs can achieve global consensus. Lastly, some simulation results are proposed to validate the effectiveness of the Q-learning algorithm and show the effect on the rate of convergence from the initial states of agents and the input saturation limit.

[1]  James Lam,et al.  Semi-Global Leader-Following Consensus of Linear Multi-Agent Systems With Input Saturation via Low Gain Feedback , 2013, IEEE Transactions on Circuits and Systems I: Regular Papers.

[2]  Bo Liu,et al.  Controllability of heterogeneous multiagent systems with two-time-scale feature. , 2019, Chaos.

[3]  Frank L. Lewis,et al.  Output synchronization of heterogeneous discrete-time systems: A model-free optimal approach , 2017, Autom..

[4]  F. Lewis,et al.  Reinforcement Learning and Feedback Control: Using Natural Decision Methods to Design Optimal Adaptive Controllers , 2012, IEEE Control Systems.

[5]  James Lam,et al.  Positive Edge-Consensus for Nodal Networks via Output Feedback , 2019, IEEE Transactions on Automatic Control.

[6]  Xin Wang,et al.  Self-triggered leader-following consensus of multi-agent systems with input time delay , 2019, Neurocomputing.

[7]  Xin Wang,et al.  Consensus of hybrid multi-agent systems by event-triggered/self-triggered strategy , 2019, Appl. Math. Comput..

[8]  Bo Liu,et al.  Second-order controllability of two-time-scale multi-agent systems , 2019, Appl. Math. Comput..

[9]  Yifan Liu,et al.  Containment control of second-order multi-agent systems via intermittent sampled position data communication , 2019, Appl. Math. Comput..

[10]  Housheng Su,et al.  Time-varying formation for linear multi-agent systems based on sampled data with multiple leaders , 2019, Neurocomputing.

[11]  Guanrong Chen,et al.  Synchronization of multi-agent systems with metric-topological interactions. , 2016, Chaos.

[12]  Housheng Su,et al.  Semi-Global Output Consensus for Discrete-Time Switching Networked Systems Subject to Input Saturation and External Disturbances , 2019, IEEE Transactions on Cybernetics.

[13]  Xia Chen,et al.  Positive Edge Consensus of Complex Networks , 2018, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[14]  Yu Zhao,et al.  Appointed-time consensus: Accurate and practical designs , 2018, Autom..

[15]  Zongli Lin,et al.  An iterative Q‐learning scheme for the global stabilization of discrete‐time linear systems subject to actuator saturation , 2019 .

[16]  Housheng Su,et al.  Discrete-Time Positive Edge-Consensus for Undirected and Directed Nodal Networks , 2018, IEEE Transactions on Circuits and Systems II: Express Briefs.

[17]  Guanrong Chen,et al.  Stabilizing Solution and Parameter Dependence of Modified Algebraic Riccati Equation With Application to Discrete-Time Network Synchronization , 2016, IEEE Transactions on Automatic Control.

[18]  Tomas Landelius,et al.  Reinforcement Learning and Distributed Local Model Synthesis , 1997 .

[19]  Xia Chen,et al.  Observer-Based Discrete-Time Nonnegative Edge Synchronization of Networked Systems , 2017, IEEE Transactions on Neural Networks and Learning Systems.

[20]  James MacLaurin,et al.  Synchronization of stochastic hybrid oscillators driven by a common switching environment. , 2018, Chaos.

[21]  Frank L. Lewis,et al.  Discrete-time dynamic graphical games: model-free reinforcement learning solution , 2015 .

[22]  Housheng Su,et al.  Observer-Based Robust Coordinated Control of Multiagent Systems With Input Saturation , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[23]  Tao Feng,et al.  Distributed Optimal Consensus Control for Nonlinear Multiagent System With Unknown Dynamic , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[24]  Hai-Tao Zhang,et al.  Swarming behaviors in multi-agent systems with nonlinear dynamics. , 2013, Chaos.

[25]  Guanrong Chen,et al.  Fully Distributed Event-Triggered Semiglobal Consensus of Multi-agent Systems With Input Saturation , 2017, IEEE Transactions on Industrial Electronics.

[26]  Housheng Su,et al.  Observer-Based Consensus for Positive Multiagent Systems With Directed Topology and Nonlinear Control Input , 2019, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[27]  Zhanshan Wang,et al.  Data-Based Optimal Control of Multiagent Systems: A Reinforcement Learning Design Approach , 2017, IEEE Transactions on Cybernetics.