A fuzzy-inference-based reinforcement learning method of overtaking decision making for automated vehicles

Intelligent decision control is one key issue of automated vehicles. Complex dynamic traffic flow and multi-requirement of passengers including vehicle safety, comfort, vehicle efficiency bring about tremendous challenges to vehicle decision making. Overtaking maneuver is a demanding task due to its large potential of traffic collision. Therefore, this paper proposes a fuzzy-inference-based reinforcement learning (FIRL) approach of autonomous overtaking decision making. Firstly, the problem of overtaking is formulated as a multi-objective Markov decision process (MDP) considering vehicle safety, driving comfort, and vehicle efficiency. Secondly, a temporal difference learning based on dynamic fuzzy (DF-TDL) is presented to learn optimized policies for autonomous overtaking decision making. Fuzzy inference is introduced to deal with continuous states and boost learning process. The RL algorithm decides whether to overtake or not based on the learned policies. Then, the automated vehicle executes local path planning and tracking. Furthermore, a simulation platform based on simulation of urban mobility (SUMO) is established to generate the random training data, that is, various traffic flows for algorithm iterative learning and validate the proposed method, extensive test results demonstrate the effectiveness of the overtaking decision-making method.

[1]  Francesco Borrelli,et al.  Predictive Active Steering Control for Autonomous Vehicle Systems , 2007, IEEE Transactions on Control Systems Technology.

[2]  Dongpu Cao,et al.  Driver Activity Recognition for Intelligent Vehicles: A Deep Learning Approach , 2019, IEEE Transactions on Vehicular Technology.

[3]  Amir Khajepour,et al.  A Potential Field-Based Model Predictive Path-Planning Controller for Autonomous Road Vehicles , 2017, IEEE Transactions on Intelligent Transportation Systems.

[4]  Yanjun Huang,et al.  Path Planning and Tracking for Vehicle Collision Avoidance Based on Model Predictive Control With Multiconstraints , 2017, IEEE Transactions on Vehicular Technology.

[5]  Hong Sun,et al.  A Novel Fuzzy Observer-Based Steering Control Approach for Path Tracking in Autonomous Vehicles , 2019, IEEE Transactions on Fuzzy Systems.

[6]  N. H. C. Yung,et al.  A Multiple-Goal Reinforcement Learning Method for Complex Vehicle Overtaking Maneuvers , 2011, IEEE Transactions on Intelligent Transportation Systems.

[7]  Long Chen,et al.  Dynamic path planning for autonomous driving on various roads with avoidance of static and moving obstacles , 2018 .

[8]  Xin Li,et al.  A Reinforcement Learning Approach to Autonomous Decision Making of Intelligent Vehicles on Highways , 2020, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[9]  Intelligent decision making for overtaking maneuver using mixed observable Markov decision process , 2018, J. Intell. Transp. Syst..

[10]  Yunfeng Ai,et al.  Driver Lane Change Intention Inference for Intelligent Vehicles: Framework, Survey, and Challenges , 2019, IEEE Transactions on Vehicular Technology.

[11]  Fei-Yue Wang,et al.  Capturing Car-Following Behaviors by Deep Learning , 2018, IEEE Transactions on Intelligent Transportation Systems.

[12]  Ching-Yao Chan,et al.  A Reinforcement Learning Based Approach for Automated Lane Change Maneuvers , 2018, 2018 IEEE Intelligent Vehicles Symposium (IV).

[13]  Xia Su,et al.  Path planning of automated guided vehicles based on improved A-Star algorithm , 2015, 2015 IEEE International Conference on Information and Automation.

[14]  Rongrong Wang,et al.  Robust H∞ output-feedback control for path following of autonomous ground vehicles , 2016 .

[15]  Julius Ziegler,et al.  Optimal trajectory generation for dynamic street scenarios in a Frenét Frame , 2010, 2010 IEEE International Conference on Robotics and Automation.

[16]  Yuanqing Xia,et al.  Lateral Path Tracking Control of Autonomous Land Vehicle Based on ADRC and Differential Flatness , 2016, IEEE Transactions on Industrial Electronics.

[17]  Hamid Reza Karimi,et al.  Finite-Time L2-Gain Asynchronous Control for Continuous-Time Positive Hidden Markov Jump Systems via T–S Fuzzy Model Approach , 2020, IEEE Transactions on Cybernetics.

[18]  Junmin Wang,et al.  Vehicle Path-Tracking Linear-Time-Varying Model Predictive Control Controller Parameter Selection Considering Central Process Unit Computational Load , 2019, Journal of Dynamic Systems, Measurement, and Control.

[19]  Hamid Reza Karimi,et al.  Fuzzy control for Electric Power Steering System with assist motor current input constraints , 2015, J. Frankl. Inst..

[20]  Fei Liu,et al.  Dynamic Self-Triggered Controller Codesign for Markov Jump Systems , 2021, IEEE Transactions on Automatic Control.

[21]  Xiang Changle,et al.  Efficient energy management strategy for hybrid electric vehicles/plug‐in hybrid electric vehicles: review and recent advances under intelligent transportation system , 2020, IET Intelligent Transport Systems.

[22]  John N. Tsitsiklis,et al.  Analysis of temporal-difference learning with function approximation , 1996, NIPS 1996.

[23]  Yun-Pang Flötteröd,et al.  Microscopic Traffic Simulation using SUMO , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[24]  Ming Yang,et al.  Conflict-Probability-Estimation-Based Overtaking for Intelligent Vehicles , 2009, IEEE Transactions on Intelligent Transportation Systems.

[25]  Dongpu Cao,et al.  Reinforcement Learning Optimized Look-Ahead Energy Management of a Parallel Hybrid Electric Vehicle , 2017, IEEE/ASME Transactions on Mechatronics.

[26]  Chao Shen,et al.  Integrated Path Planning and Tracking Control of an AUV: A Unified Receding Horizon Optimization Approach , 2017, IEEE/ASME Transactions on Mechatronics.