Learning Stabilizable Dynamical Systems via Control Contraction Metrics

We propose a novel framework for learning stabilizable nonlinear dynamical systems for continuous control tasks in robotics. The key idea is to develop a new control-theoretic regularizer for dynamics fitting rooted in the notion of stabilizability, which guarantees that the learned system can be accompanied by a robust controller capable of stabilizing any open-loop trajectory that the system may generate. By leveraging tools from contraction theory, statistical learning, and convex optimization, we provide a general and tractable semi-supervised algorithm to learn stabilizable dynamics, which can be applied to complex underactuated systems. We validated the proposed algorithm on a simulated planar quadrotor system and observed notably improved trajectory generation and tracking performance with the control-theoretic regularized model over models learned using traditional regression techniques, especially when using a small number of demonstration examples. The results presented illustrate the need to infuse standard model-based reinforcement learning algorithms with concepts drawn from nonlinear control theory for improved reliability.

[1]  I. Michael Ross,et al.  Direct Trajectory Optimization by a Chebyshev Pseudospectral Method ; Journal of Guidance, Control, and Dynamics, v. 25, 2002 ; pp. 160-166 , 2002 .

[2]  Ashwin P. Dani,et al.  Learning Partially Contracting Dynamical Systems from Demonstrations , 2017, CoRL.

[3]  Aude Billard,et al.  Learning Stable Nonlinear Dynamical Systems With Gaussian Mixture Models , 2011, IEEE Transactions on Robotics.

[4]  J. Slotine,et al.  On the Adaptive Control of Robot Manipulators , 1987 .

[5]  Martial Hebert,et al.  Improving Multi-Step Prediction of Learned Time Series Models , 2015, AAAI.

[6]  Oussama Khatib,et al.  Learning potential functions from human demonstrations with encapsulated dynamic and compliant behaviors , 2017, Auton. Robots.

[7]  Carl E. Rasmussen,et al.  PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.

[8]  Martial Hebert,et al.  Improved Learning of Dynamics Models for Control , 2016, ISER.

[9]  Andreas Krause,et al.  Safe Model-based Reinforcement Learning with Stability Guarantees , 2017, NIPS.

[10]  Tengyuan Liang,et al.  Just Interpolate: Kernel "Ridgeless" Regression Can Generalize , 2018, The Annals of Statistics.

[11]  Ian R. Manchester,et al.  An Amendment to "Control Contraction Metrics: Convex and Intrinsic Criteria for Nonlinear Feedback Design" , 2017, ArXiv.

[12]  Marco Pavone,et al.  Learning Stabilizable Dynamical Systems via Control Contraction Metrics , 2018, WAFR.

[13]  Jean-Jacques E. Slotine,et al.  On Contraction Analysis for Non-linear Systems , 1998, Autom..

[14]  Klaus Neumann,et al.  Neural learning of vector fields for encoding stable dynamical systems , 2014, Neurocomputing.

[15]  Sergey Levine,et al.  Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[16]  Marco Pavone,et al.  Robust online motion planning via contraction theory and convex optimization , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[17]  Robert M. Sanner,et al.  Gaussian Networks for Direct Adaptive Control , 1991, 1991 American Control Conference.

[18]  Kenneth O. Kortanek,et al.  Semi-Infinite Programming: Theory, Methods, and Applications , 1993, SIAM Rev..

[19]  Marco A. López,et al.  A New Exchange Method for Convex Semi-Infinite Programming , 2010, SIAM J. Optim..

[20]  A. Schaft,et al.  Variational and Hamiltonian Control Systems , 1987 .

[21]  Aude Billard,et al.  Learning Stable Task Sequences from Demonstration with Linear Parameter Varying Systems and Hidden Markov Models , 2017, CoRL.

[22]  Ian R. Manchester,et al.  Robust Control Contraction Metrics: A Convex Approach to Nonlinear State-Feedback ${H}^\infty$ Control , 2018, IEEE Control Systems Letters.