Aggressive Online Control of a Quadrotor via Deep Network Representations of Optimality Principles

Optimal control holds great potential to improve a variety of robotic applications. The application of optimal control on-board limited platforms has been severely hindered by the large computational requirements of current state of the art implementations. In this work, we make use of a deep neural network to directly map the robot states to control actions. The network is trained offline to imitate the optimal control computed by a time consuming direct nonlinear method. A mixture of time optimality and power optimality is considered with a continuation parameter used to select the predominance of each objective. We apply our networks (termed G&CNets) to aggressive quadrotor control, first in simulation and then in the real world. We give insight into the factors that influence the ‘reality gap’ between the quadrotor model used by the offline optimal control method and the real quadrotor. Furthermore, we explain how we set up the model and the control structure on-board of the real quadrotor to successfully close this gap and perform time-optimal maneuvers in the real world. Finally, G&CNet’s performance is compared to state-of-the-art differential-flatness-based optimal control methods. We show, in the experiments, that G&CNets lead to significantly faster trajectory execution due to, in part, the less restrictive nature of the allowed state-to-input mappings.

[1]  Raffaello D'Andrea,et al.  Quadrocopter Trajectory Generation and Control , 2011 .

[2]  Sergio Salazar,et al.  Optimized Discrete Control Law for Quadrotor Stabilization: Experimental Results , 2016, J. Intell. Robotic Syst..

[3]  Dario Izzo,et al.  Real-time optimal control via Deep Neural Networks: study on landing problems , 2016, ArXiv.

[4]  Dario Izzo,et al.  Learning the optimal state-feedback via supervised imitation learning , 2019, Astrodynamics.

[5]  Roland Siegwart,et al.  Control of a Quadrotor With Reinforcement Learning , 2017, IEEE Robotics and Automation Letters.

[6]  Kris Hauser,et al.  Learning Trajectories for Real- Time Optimal Control of Quadrotors , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[7]  Vijay Kumar,et al.  Minimum snap trajectory generation and control for quadrotors , 2011, 2011 IEEE International Conference on Robotics and Automation.

[8]  Fabio Morbidi,et al.  Minimum-energy path generation for a quadrotor UAV , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[9]  Charles Richter,et al.  Aggressive flight of fixed-wing and quadrotor aircraft in dense indoor environments , 2015, Int. J. Robotics Res..

[10]  Raffaello D'Andrea,et al.  Performance benchmarking of quadrotor systems using time-optimal control , 2012, Auton. Robots.

[11]  Yaonan Wang,et al.  Adaptive RBFNNs/integral sliding mode control for a quadrotor aircraft , 2016, Neurocomputing.

[12]  Antonio Franchi,et al.  Differential Flatness of Quadrotor Dynamics Subject to Rotor Drag for Accurate Tracking of High-Speed Trajectories , 2017, IEEE Robotics and Automation Letters.

[13]  Nicolas Mansard,et al.  Trajectory generation for quadrotor based systems using numerical optimal control , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[14]  Sertac Karaman,et al.  Accurate Tracking of Aggressive Quadrotor Trajectories Using Incremental Nonlinear Dynamic Inversion and Differential Flatness , 2018, 2018 IEEE Conference on Decision and Control (CDC).

[15]  Guido C. H. E. de Croon,et al.  Adaptive Incremental Nonlinear Dynamic Inversion for Attitude Control of Micro Air Vehicles , 2016 .

[16]  Angela P. Schoellig,et al.  Deep neural networks for improved, impromptu trajectory tracking of quadrotors , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[17]  Dario Izzo,et al.  On the Stability Analysis of Deep Neural Network Representations of an Optimal State Feedback , 2018, IEEE Transactions on Aerospace and Electronic Systems.