Hierarchical Evasive Path Planning Using Reinforcement Learning and Model Predictive Control

Motion planning plays an essential role in designing self-driving functions for connected and autonomous vehicles. The methods need to provide a feasible trajectory for the vehicle to follow, fulfilling different requirements, such as safety, efficiency, and passenger comfort. In this area, algorithms must also meet strict real-time expectations, since, especially in an emergency, the decision time is limited, which raises a trade-off for the feasibility requirements. This article proposes a hierarchical path planning solution for evasive maneuvering, where a Twin Delayed DDPG reinforcement learning agent generates the parameters of a geometric path consisting of chlotoids and straight sections, and an underlying model predictive control loop fulfills the trajectory following tasks. The method is applied to the automotive double lane-change test, a common emergency situation, comparing its results with human drivers’ performance using a dynamic simulation environment. Besides the test’s standardized parameters, a broader range of topological layouts is chosen, both for the training and performance evaluation. The results show that the proposed method highly outperforms human drivers, especially in challenging situations, while meeting the computational requirements, as the pre-trained neural network and path generation algorithm can provide a solution in an instant, based on the experience gained during the training process.

[1]  Hans B. Pacejka Applications of Transient Tire Models , 2012 .

[2]  Guy Lever,et al.  Deterministic Policy Gradient Algorithms , 2014, ICML.

[3]  Christof Büskens,et al.  Controlling an Autonomous Vehicle with Deep Reinforcement Learning , 2019, 2019 IEEE Intelligent Vehicles Symposium (IV).

[4]  István Varga,et al.  Development of a Test Track for Driverless Cars: Vehicle Design, Track Configuration, and Liability Considerations , 2017 .

[5]  Xiaohui Zhang,et al.  Automated vehicle's behavior decision making using deep reinforcement learning and high-fidelity simulation environment , 2018, Transportation Research Part C: Emerging Technologies.

[6]  Ali Ghaffari,et al.  Modeling of Double Lane Change Maneuver of Vehicles , 2018 .

[7]  Sebastian Thrun,et al.  Junior: The Stanford entry in the Urban Challenge , 2008, J. Field Robotics.

[8]  Florian Kuhnt,et al.  Adaptive Behavior Generation for Autonomous Driving using Deep Reinforcement Learning with Compact Semantic States , 2018, 2018 IEEE Intelligent Vehicles Symposium (IV).

[9]  Szil'ard Aradi,et al.  Survey of Deep Reinforcement Learning for Motion Planning of Autonomous Vehicles , 2020, IEEE Transactions on Intelligent Transportation Systems.

[10]  Germán Ros,et al.  CARLA: An Open Urban Driving Simulator , 2017, CoRL.

[11]  R. Mazo On the theory of brownian motion , 1973 .

[12]  Kaiming Yang,et al.  Shared Steering Torque Control for Lane Change Assistance: A Stochastic Game-Theoretic Approach , 2019, IEEE Transactions on Industrial Electronics.

[13]  Hyuk-Jun Chang,et al.  Explicit model predictive control for linear time-variant systems with application to double-lane-change maneuver , 2018, PloS one.

[14]  Giulio Panzani,et al.  H∞ control with look-ahead for lane keeping in autonomous vehicles , 2017, 2017 IEEE Conference on Control Technology and Applications (CCTA).

[15]  Carl-Johan Hoel,et al.  Automated Speed and Lane Change Decision Making using Deep Reinforcement Learning , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[16]  P. Gáspár,et al.  Motion Planning for Highly Automated Road Vehicles with a Hybrid Approach Using Nonlinear Optimization and Artificial Neural Networks , 2019, Strojniški vestnik – Journal of Mechanical Engineering.

[17]  Qi Zhu,et al.  A practical trajectory planning framework for autonomous ground vehicles driving in urban environments , 2015, 2015 IEEE Intelligent Vehicles Symposium (IV).

[18]  Sebastian Thrun,et al.  Anytime Dynamic A*: An Anytime, Replanning Algorithm , 2005, ICAPS.

[19]  Baopu Li,et al.  A Control Strategy of Autonomous Vehicles Based on Deep Reinforcement Learning , 2016, 2016 9th International Symposium on Computational Intelligence and Design (ISCID).

[20]  Xinkai Wu,et al.  A lane-change trajectory model from drivers’ vision view , 2017 .

[21]  Zsolt Szalay,et al.  Conflicts of Automated Driving With Conventional Traffic Infrastructure , 2020, IEEE Access.

[22]  Moulay A. Akhloufi,et al.  Learning to Drive by Imitation: An Overview of Deep Behavior Cloning Methods , 2021, IEEE Transactions on Intelligent Vehicles.

[23]  Guodong Yin,et al.  Active collision algorithm for autonomous electric vehicles at intersections , 2019 .

[24]  H. Jin Kim,et al.  Autonomous lane keeping based on approximate Q-learning , 2017, 2017 14th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI).

[25]  Vicenç Puig,et al.  LPV-MP planning for autonomous racing vehicles considering obstacles , 2020, Robotics Auton. Syst..

[26]  Rajesh Rajamani,et al.  Vehicle dynamics and control , 2005 .

[27]  B. R. Mehta,et al.  Advanced process control systems , 2015 .

[28]  Péter Gáspár,et al.  Proving Ground Test of a DDPG-based Vehicle Trajectory Planner , 2020, 2020 European Control Conference (ECC).

[29]  Dingli Yu,et al.  The indirect shared steering control under double loop structure of driver and automation , 2020, IEEE/CAA Journal of Automatica Sinica.

[30]  Doran Wilde Computing clothoid segments for trajectory generation , 2009, 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[31]  Yosuke Nishimura,et al.  Application of Stackelberg Game Theory for Shared Steering Torque Control in Lane Change Maneuver , 2018, 2018 IEEE Intelligent Vehicles Symposium (IV).

[32]  Fazel Naghdy,et al.  Model predictive control-based lane change control system for an autonomous vehicle , 2016, 2016 IEEE Region 10 Conference (TENCON).

[33]  D. Ferguson,et al.  Motion planning in urban environments: Part II , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[34]  Johann Marius Zöllner,et al.  Learning how to drive in a real world simulation with deep Q-Networks , 2017, 2017 IEEE Intelligent Vehicles Symposium (IV).

[35]  Tamás Bécsi,et al.  Model Based Trajectory Planning for Highly Automated Road Vehicles , 2017 .

[36]  Junjie Wang,et al.  Lane Change Decision-making through Deep Reinforcement Learning with Rule-based Constraints , 2019, 2019 International Joint Conference on Neural Networks (IJCNN).

[37]  Nasser L. Azad,et al.  Longitudinal Dynamic versus Kinematic Models for Car-Following Control Using Deep Reinforcement Learning , 2019, 2019 IEEE Intelligent Transportation Systems Conference (ITSC).

[38]  Marco Forgione,et al.  Efficient Calibration of Embedded MPC , 2019, ArXiv.

[39]  Yujun Zeng,et al.  Deep, Consistent Behavioral Decision Making with Planning Features for Autonomous Vehicles , 2019, Electronics.

[40]  Duan Jianmin,et al.  Optimal lane change motion of intelligent vehicles based on extended adaptive pseudo-spectral method under uncertain vehicle mass , 2017 .

[41]  Saïd Mammar,et al.  Geometric continuous-curvature path planning for automatic parallel parking , 2013, 2013 10th IEEE INTERNATIONAL CONFERENCE ON NETWORKING, SENSING AND CONTROL (ICNSC).

[42]  Emilio Frazzoli,et al.  A Survey of Motion Planning and Control Techniques for Self-Driving Urban Vehicles , 2016, IEEE Transactions on Intelligent Vehicles.

[43]  Dominique Gruyer,et al.  A Review of Motion Planning for Highway Autonomous Driving , 2020, IEEE Transactions on Intelligent Transportation Systems.

[44]  Péter Gáspár,et al.  Hybrid DDPG Approach for Vehicle Motion Planning , 2019, ICINCO.

[45]  A. Lambert,et al.  Path Planning using a Dynamic Vehicle Model , 2006, 2006 2nd International Conference on Information & Communication Technologies.

[46]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[47]  Jonathan P. How,et al.  Real-Time Motion Planning With Applications to Autonomous Urban Driving , 2009, IEEE Transactions on Control Systems Technology.

[48]  Yuan Zhu,et al.  Deep Q-Network Based Decision Making for Autonomous Driving , 2019, 2019 3rd International Conference on Robotics and Automation Sciences (ICRAS).

[49]  Yuval Tassa,et al.  Continuous control with deep reinforcement learning , 2015, ICLR.