Longitudinal deep truck: Deep longitudinal model with application to sim2real deep reinforcement learning for heavy‐duty truck control in the field

To develop cooperative adaptive cruise control (CACC), the choice of control approach often influences and can limit the choice of model structure, and vice versa. For heavy‐duty trucks, practical application of CACC in the field is heavily influenced by the accuracy of the used model. Deep learning and deep reinforcement learning (deep‐RL) have recently been used to demonstrate improved modeling and control performance for vehicles such as cars and quadrotors compared to state‐of‐the‐art. The literature on the application of deep learning and deep‐RL for heavy‐duty trucks in the field, which are significantly more complex than cars, is still sparse, however. In this article, we develop a two‐layer gray‐box deep learning model to capture longitudinal dynamics of heavy‐duty trucks while abstracting their complexity and present an approach to properly break the nested feedback loops in the model for training. We compare this model with three other alternative models and show that it achieves ~ 10 x $\unicode{x0007E}10x$ better general performance compared to a standard artificial neural network and results in ~ 4 x $\unicode{x0007E}4x$ and ~ 40 x $\unicode{x0007E}40x$ slower steady‐state acceleration and speed error growth rates, respectively. We then present an architecture to utilize these deep learning models within the deep‐RL framework and use it to develop baseline CACC controllers that can be zero‐shot transferred to the field. To carry out the work, we present a setup of differently configured trucks along with their interface architecture and stochastic driving cycle generators for data collection. Numerical validation of the approach demonstrated stationary and bounded modeling error, and demonstrated transfer of CACC controllers with consistent overshoot bounds and a stable approximately‐zero steady‐state error. Validation from field experiments demonstrated similarly consistent results. Compared to a state‐of‐the‐art benchmark, the deep‐RL controller achieved lower speed and time‐gap error variance but higher time‐gap error offset.

[1]  Mengchu Zhou,et al.  Solving Dynamic Traveling Salesman Problems With Deep Reinforcement Learning , 2021, IEEE Transactions on Neural Networks and Learning Systems.

[2]  A. Bayen,et al.  Deep truck cruise control , 2022, Control Engineering Practice.

[3]  E. Gavves,et al.  Domain- and task-specific transfer learning for medical segmentation tasks , 2021, Comput. Methods Programs Biomed..

[4]  Derek Rayside,et al.  A sensorless state estimation for a safety-oriented cyber-physical system in urban driving: Deep learning approach , 2021, IEEE/CAA Journal of Automatica Sinica.

[5]  Eric Medvet,et al.  Crossing the Reality Gap: a Survey on Sim-to-Real Transferability of Robot Controllers in Reinforcement Learning , 2021, IEEE Access.

[6]  Tomi Westerlund,et al.  Sim-to-Real Transfer in Deep Reinforcement Learning for Robotics: a Survey , 2020, 2020 IEEE Symposium Series on Computational Intelligence (SSCI).

[7]  Sean R. Anderson,et al.  Longitudinal Vehicle Dynamics: A Comparison of Physical and Data-Driven Models Under Large-Scale Real-World Driving Conditions , 2020, IEEE Access.

[8]  Feiyue Wang,et al.  Parallel distance: A new paradigm of measurement for parallel driving , 2020, IEEE/CAA Journal of Automatica Sinica.

[9]  Shuo Wang,et al.  Real–Sim–Real Transfer for Real-World Robot Control Policy Learning with Deep Reinforcement Learning , 2020, Applied Sciences.

[10]  Daniele Bortoluzzi,et al.  Modelling longitudinal vehicle dynamics with neural networks , 2020, Vehicle System Dynamics.

[11]  Alexandre M. Bayen,et al.  Deep Truck : A deep neural network model for longitudinal dynamics of heavy duty trucks , 2019, 2019 IEEE Intelligent Transportation Systems Conference (ITSC).

[12]  J. Christian Gerdes,et al.  Neural network vehicle models for high-performance automated driving , 2019, Science Robotics.

[13]  Jaime F. Fisac,et al.  A General Safety Framework for Learning-Based Control in Uncertain Robotic Systems , 2017, IEEE Transactions on Automatic Control.

[14]  Hong Wang,et al.  Parallel planning: a new motion planning framework for autonomous driving , 2019, IEEE/CAA Journal of Automatica Sinica.

[15]  Sen Wang,et al.  Deep Reinforcement Learning for Autonomous Driving , 2018, ArXiv.

[16]  Fawzi Nashashibi,et al.  End-to-End Race Driving with Deep Reinforcement Learning , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[17]  Steven Lake Waslander,et al.  Deep Learning a Quadrotor Dynamic Model for Multi-Step Prediction , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[18]  Xiao-Yun Lu,et al.  Influences on Energy Savings of Heavy Trucks Using Cooperative Adaptive Cruise Control , 2018 .

[19]  Alexandre M. Bayen,et al.  Expert Level Control of Ramp Metering Based on Multi-Task Deep Reinforcement Learning , 2017, IEEE Transactions on Intelligent Transportation Systems.

[20]  J. C. Gerdes,et al.  Tire Modeling to Enable Model Predictive Control of Automated Vehicles From Standstill to the Limits of Handling , 2018 .

[21]  Jan Peters,et al.  Learning inverse dynamics models in O(n) time with LSTM networks , 2017, 2017 IEEE-RAS 17th International Conference on Humanoid Robotics (Humanoids).

[22]  Alexandre M. Bayen,et al.  Flow: Architecture and Benchmarking for Reinforcement Learning in Traffic Control , 2017, ArXiv.

[23]  Xiao-Yun Lu,et al.  Integrated ACC and CACC development for Heavy-Duty Truck partial automation , 2017, 2017 American Control Conference (ACC).

[24]  Yu Wang,et al.  A new concept using LSTM Neural Networks for dynamic system identification , 2017, 2017 American Control Conference (ACC).

[25]  Wojciech Zaremba,et al.  Domain randomization for transferring deep neural networks from simulation to the real world , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[26]  Sergey Levine,et al.  Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[27]  Karl Henrik Johansson,et al.  Cooperative Look-Ahead Control for Fuel-Efficient and Safe Heavy-Duty Vehicle Platooning , 2015, IEEE Transactions on Control Systems Technology.

[28]  Claire J. Tomlin,et al.  Learning quadrotor dynamics using neural network for flight control , 2016, 2016 IEEE 55th Conference on Decision and Control (CDC).

[29]  Razvan Pascanu,et al.  Progressive Neural Networks , 2016, ArXiv.

[30]  Pieter Abbeel,et al.  Benchmarking Deep Reinforcement Learning for Continuous Control , 2016, ICML.

[31]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[32]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[33]  Sergey Levine,et al.  Trust Region Policy Optimization , 2015, ICML.

[34]  Christopher Nowakowski,et al.  Cooperative Adaptive Cruise Control , 2015 .

[35]  Xiao-Yun Lu,et al.  COOPERATIVE ADAPTIVE CRUISE CONTROL (CACC) DEFINITIONS AND OPERATING CONCEPTS , 2015 .

[36]  Vicente Milanés Montero,et al.  Cooperative Adaptive Cruise Control in Real Traffic Situations , 2014, IEEE Transactions on Intelligent Transportation Systems.

[37]  Johannes P. Schlöder,et al.  Mixed-integer NMPC for predictive cruise control of heavy-duty trucks , 2013, 2013 European Control Conference (ECC).

[38]  Steven E Shladover,et al.  Impacts of Cooperative Adaptive Cruise Control on Freeway Traffic Flow , 2012 .

[39]  Yoram Singer,et al.  Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..

[40]  Matthew J. Streeter,et al.  Adaptive Bound Optimization for Online Convex Optimization , 2010, COLT 2010.

[41]  van de Mjg René Molengraft,et al.  Cooperative adaptive cruise control , 2009 .

[42]  Luke Ng,et al.  Reinforcement Learning of Dynamic Collaborative Driving , 2008 .

[43]  Francesco Borrelli,et al.  MPC-Based Approach to Active Steering for Autonomous Vehicle Systems , 2005 .

[44]  J. Karl Hedrick,et al.  Heavy-duty vehicle modelling and longitudinal control , 2005 .

[45]  Thomas Connolly,et al.  The Predictive Cruise Control – A System to Reduce Fuel Consumption of Heavy Duty Trucks , 2004 .

[46]  Anna G. Stefanopoulou,et al.  Simultaneous mass and time-varying grade estimation for heavy-duty vehicles , 2003, Proceedings of the 2003 American Control Conference, 2003..

[47]  J. Christian Gerdes,et al.  Parameter Estimation and Command Modification for Longitudinal Control of Heavy Vehicles , 2003 .

[48]  Anna G. Stefanopoulou,et al.  Speed gradient approach to longitudinal control of heavy-duty vehicles equipped with variable compression brake , 2002, IEEE Trans. Control. Syst. Technol..

[49]  Michael W. Sayers,et al.  Modeling assumptions for realistic multibody simulations of the yaw and roll behavior of heavy trucks , 1996 .