Model-Based Actor-Critic Learning for Optimal Tracking Control of Robots With Input Saturation

[1]  Ning Sun,et al.  Adaptive Control for Pneumatic Artificial Muscle Systems With Parametric Uncertainties and Unidirectional Input Constraints , 2020, IEEE Transactions on Industrial Informatics.

[2]  Mohammad Emtiyaz Khan,et al.  TD-regularized actor-critic methods , 2018, Machine Learning.

[3]  Takayuki Kanda,et al.  Robot behavior adaptation for human-robot interaction based on policy gradient reinforcement learning , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[4]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[5]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[6]  S.H.G. ten Hagen,et al.  Linear Quadratic Regulation using reinforcement learning , 1998 .

[7]  Jun Wang,et al.  A dual neural network for redundancy resolution of kinematically redundant manipulators subject to joint limits and joint velocity limits , 2003, IEEE Trans. Neural Networks.

[8]  Frank L. Lewis,et al.  Discrete-Time Nonlinear HJB Solution Using Approximate Dynamic Programming: Convergence Proof , 2008, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[9]  Shuai Li,et al.  Manipulability Optimization of Redundant Manipulators Using Dynamic Neural Networks , 2017, IEEE Transactions on Industrial Electronics.

[10]  Paolo Rocco,et al.  Trajectory Generation for Unmanned Aerial Manipulators Through Quadratic Programming , 2017, IEEE Robotics and Automation Letters.

[11]  Frank L. Lewis,et al.  Reinforcement Q-learning for optimal tracking control of linear discrete-time systems with unknown dynamics , 2014, Autom..

[12]  Stefan Schaal,et al.  Natural Actor-Critic , 2003, Neurocomputing.

[13]  Goro Obinata,et al.  Development of a Wearable Robot for Assisting Carpentry Workers , 2006 .

[14]  Frank L. Lewis,et al.  Model-free Q-learning designs for linear discrete-time zero-sum games with application to H-infinity control , 2007, Autom..

[15]  Richard S. Sutton,et al.  A Menu of Designs for Reinforcement Learning Over Time , 1995 .

[16]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.

[17]  Huaguang Zhang,et al.  Adaptive Dynamic Programming: An Introduction , 2009, IEEE Computational Intelligence Magazine.

[18]  Sergey Levine,et al.  Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[19]  Stefan Schaal,et al.  Policy Gradient Methods for Robotics , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[20]  Shin'ichi Yuta,et al.  Baggage Transportation and Navigation by a Wheeled Inverted Pendulum Mobile Robot , 2009, IEEE Transactions on Industrial Electronics.

[21]  Bolin Liao,et al.  Recurrent Neural Network for Kinematic Control of Redundant Manipulators With Periodic Input Disturbance and Physical Constraints , 2019, IEEE Transactions on Cybernetics.

[22]  Shin Ishii,et al.  Reinforcement Learning for CPG-Driven Biped Robot , 2004, AAAI.

[23]  Sarangapani Jagannathan,et al.  Optimal tracking control of affine nonlinear discrete-time systems with unknown internal dynamics , 2009, Proceedings of the 48h IEEE Conference on Decision and Control (CDC) held jointly with 2009 28th Chinese Control Conference.

[24]  Guy Lever,et al.  Deterministic Policy Gradient Algorithms , 2014, ICML.

[25]  S. Shankar Sastry,et al.  Autonomous Helicopter Flight via Reinforcement Learning , 2003, NIPS.

[26]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[27]  Yuval Tassa,et al.  Continuous control with deep reinforcement learning , 2015, ICLR.

[28]  V. Gullapalli,et al.  Acquiring robot skills via reinforcement learning , 1994, IEEE Control Systems.

[29]  Kejia Zhuang,et al.  Analysis and assessment of robotic belt grinding mechanisms by force modeling and force control experiments , 2018 .

[30]  He Chen,et al.  Neural Network-Based Adaptive Antiswing Control of an Underactuated Ship-Mounted Crane With Roll Motions and Input Dead Zones , 2020, IEEE Transactions on Neural Networks and Learning Systems.

[31]  Tadej Petric,et al.  Human-in-the-loop approach for teaching robot assembly tasks using impedance control interface , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[32]  Huaguang Zhang,et al.  A Novel Infinite-Time Optimal Tracking Control Scheme for a Class of Discrete-Time Nonlinear Systems via the Greedy HDP Iteration Algorithm , 2008, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[33]  Judy A. Franklin,et al.  Biped dynamic walking using reinforcement learning , 1997, Robotics Auton. Syst..