Data-Driven Model Predictive Control for Trajectory Tracking With a Robotic Arm

High-precision trajectory tracking is fundamental in robotic manipulation. While industrial robots address this through stiffness and high-performance hardware, compliant and cost-effective robots require advanced control to achieve accurate position tracking. In this letter, we present a model-based control approach, which makes use of data gathered during operation to improve the model of the robotic arm and thereby the tracking performance. The proposed scheme is based on an inverse dynamics feedback linearization and a data-driven error model, which are integrated into a model predictive control formulation. In particular, we show how offset-free tracking can be achieved by augmenting a nominal model with both a Gaussian process, which makes use of offline data, and an additive disturbance model suitable for efficient online estimation of the residual disturbance via an extended Kalman filter. The performance of the proposed offset-free GPMPC scheme is demonstrated on a compliant 6 degrees of freedom robotic arm, showing significant performance improvements compared to other robot control algorithms.

[1]  Jan Peters,et al.  Learning inverse dynamics models in O(n) time with LSTM networks , 2017, 2017 IEEE-RAS 17th International Conference on Humanoid Robotics (Humanoids).

[2]  J. J. Slotine,et al.  Tracking control of non-linear systems using sliding surfaces with application to robot manipulators , 1983, 1983 American Control Conference.

[3]  Alessandro Chiuso,et al.  Online semi-parametric learning for inverse dynamics modeling , 2016, 2016 IEEE 55th Conference on Decision and Control (CDC).

[4]  Lakmal Seneviratne,et al.  Adaptive Control Of Robot Manipulators , 1992, Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems.

[5]  James B. Rawlings,et al.  A new autocovariance least-squares method for estimating noise covariances , 2006, Autom..

[6]  Kim Peter Wabersich,et al.  Linear Model Predictive Safety Certification for Learning-Based Control , 2018, 2018 IEEE Conference on Decision and Control (CDC).

[7]  Athanasios S. Polydoros,et al.  Real-time deep learning of robotic manipulator inverse dynamics , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[8]  Angela P. Schoellig,et al.  Robust Constrained Learning-based NMPC enabling reliable mobile robot path tracking , 2016, Int. J. Robotics Res..

[9]  Giorgio Metta,et al.  Incremental semiparametric inverse dynamics learning , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[10]  Stefan Schaal,et al.  Reinforcement learning of motor skills in high dimensions: A path integral approach , 2010, 2010 IEEE International Conference on Robotics and Automation.

[11]  Manfred Morari,et al.  Offset free model predictive control , 2007, 2007 46th IEEE Conference on Decision and Control.

[12]  Marc Peter Deisenroth,et al.  Data-Efficient Reinforcement Learning with Probabilistic Model Predictive Control , 2017, AISTATS.

[13]  Marc Peter Deisenroth,et al.  Efficient reinforcement learning using Gaussian processes , 2010 .

[14]  Duy Nguyen-Tuong,et al.  Optimizing Long-term Predictions for Model-based Policy Search , 2017, CoRL.

[15]  Carl E. Rasmussen,et al.  Manifold Gaussian Processes for regression , 2014, 2016 International Joint Conference on Neural Networks (IJCNN).

[16]  S. Shankar Sastry,et al.  Provably safe and robust learning-based model predictive control , 2011, Autom..

[17]  Agathe Girard,et al.  Prediction at an Uncertain Input for Gaussian Processes and Relevance Vector Machines Application to Multiple-Step Ahead Time-Series Forecasting , 2002 .

[18]  Michalis K. Titsias,et al.  Variational Learning of Inducing Variables in Sparse Gaussian Processes , 2009, AISTATS.

[19]  Peter Fankhauser,et al.  ANYmal - a highly mobile and dynamic quadrupedal robot , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[20]  Malte Kuß,et al.  Gaussian process models for robust regression, classification, and reinforcement learning , 2006 .

[21]  Alexander Liniger,et al.  Cautious NMPC with Gaussian Process Dynamics for Miniature Race Cars , 2017, ArXiv.

[22]  S. Shankar Sastry,et al.  Dynamic control of redundant manipulators , 1989, J. Field Robotics.

[23]  Alexander J. Smola,et al.  Sparse Greedy Gaussian Process Regression , 2000, NIPS.

[24]  Stefan Schaal,et al.  Towards robust online inverse dynamics learning , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[25]  Bruno Siciliano,et al.  Resolved-acceleration control of robot manipulators: A critical review with experiments , 1998, Robotica.

[26]  Jan Peters,et al.  Using model knowledge for learning inverse dynamics , 2010, 2010 IEEE International Conference on Robotics and Automation.

[27]  Kevin L. Moore,et al.  Iterative Learning Control: Brief Survey and Categorization , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[28]  Takamitsu Matsubara,et al.  Reinforcement Learning Ship Autopilot: Sample-efficient and Model Predictive Control-based Approach , 2019, ArXiv.

[29]  Angela P. Schoellig,et al.  Learning-based nonlinear model predictive control to improve vision-based mobile robot path-tracking in challenging outdoor environments , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[30]  Marco Hutter,et al.  Deployment of an autonomous mobile manipulator at MBZIRC , 2018, J. Field Robotics.

[31]  Pietro Falco,et al.  Data-efficient control policy search using residual dynamics learning , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[32]  Jan M. Maciejowski,et al.  Learning-based Nonlinear Model Predictive Control , 2017 .

[33]  Ian Brooks,et al.  Mobile robotic fabrication at 1:1 scale: the In situ Fabricator , 2017, Construction Robotics.

[34]  Zoubin Ghahramani,et al.  Sparse Gaussian Processes using Pseudo-inputs , 2005, NIPS.

[35]  Markus Giftthaler,et al.  Towards a Unified Framework of Efficient Algorithms for Numerical Optimal Robot Control , 2018 .

[36]  Takamitsu Matsubara,et al.  Reinforcement Learning Boat Autopilot: A Sample-efficient and Model Predictive Control based Approach , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[37]  M. Wendt,et al.  Nonlinear Chance-Constrained Process Optimization under Uncertainty , 2002 .

[38]  Carl E. Rasmussen,et al.  PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.

[39]  Juraj Kabzan,et al.  Cautious Model Predictive Control Using Gaussian Process Regression , 2017, IEEE Transactions on Control Systems Technology.

[40]  Torsten Koller,et al.  Learning-based Model Predictive Control for Safe Exploration and Reinforcement Learning , 2019, ArXiv.

[41]  Stefan Schaal,et al.  A new data source for inverse dynamics learning , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[42]  Mark W. Spong,et al.  Comments on "Adaptive manipulator control: a case study" by J. Slotine and W. Li , 1990 .

[43]  Oussama Khatib,et al.  A unified approach for motion and force control of robot manipulators: The operational space formulation , 1987, IEEE J. Robotics Autom..

[44]  Stefan Schaal,et al.  Locally Weighted Projection Regression : An O(n) Algorithm for Incremental Real Time Learning in High Dimensional Space , 2000 .