Deep Reinforcement Learning Based Left-Turn Connected and Automated Vehicle Control at Signalized Intersection in Vehicle-to-Infrastructure Environment

In order to solve the problem of vehicle delay caused by stops at signalized intersections, a micro-control method of a left-turning connected and automated vehicle (CAV) based on an improved deep deterministic policy gradient (DDPG) is designed in this paper. In this paper, the micro-control of the whole process of a left-turn vehicle approaching, entering, and leaving a signalized intersection is considered. In addition, in order to solve the problems of low sampling efficiency and overestimation of the critic network of the DDPG algorithm, a positive and negative reward experience replay buffer sampling mechanism and multi-critic network structure are adopted in the DDPG algorithm in this paper. Finally, the effectiveness of the signal control method, six DDPG-based methods (DDPG, PNRERB-1C-DDPG, PNRERB-3C-DDPG, PNRERB-5C-DDPG, PNRERB-5CNG-DDPG, and PNRERB-7C-DDPG), and four DQN-based methods (DQN, Dueling DQN, Double DQN, and Prioritized Replay DQN) are verified under 0.2, 0.5, and 0.7 saturation degrees of left-turning vehicles at a signalized intersection within a VISSIM simulation environment. The results show that the proposed deep reinforcement learning method can get a number of stops benefits ranging from 5% to 94%, stop time benefits ranging from 1% to 99%, and delay benefits ranging from −17% to 93%, respectively compared with the traditional signal control method.

[1]  Yukimasa Matsumoto,et al.  Reinforcement Learning of Driver Receiving Traffic Signal Information for Passing through Signalized Intersection at Arterial Road , 2019, Transportation Research Procedia.

[2]  Saiedeh N. Razavi,et al.  Evaluating reinforcement learning state representations for adaptive traffic signal control , 2019, ANT/SEIT.

[3]  Ardalan Vahidi,et al.  Energy saving potentials of connected and automated vehicles , 2018, Transportation Research Part C: Emerging Technologies.

[4]  Xinkai Wu,et al.  Eco-driving advisory strategies for a platoon of mixed gasoline and electric vehicles in a connected vehicle system , 2018, Transportation Research Part D: Transport and Environment.

[5]  Markos Papageorgiou,et al.  A platoon based cooperative eco-driving model for mixed automated and human-driven vehicles at a signalised intersection , 2018, Transportation Research Part C: Emerging Technologies.

[6]  Ian H. Witten,et al.  An Adaptive Optimal Controller for Discrete-Time Markov Environments , 1977, Inf. Control..

[7]  Zhigang Zhang,et al.  Double-DQN based path smoothing and tracking control method for robotic vehicle navigation , 2019, Comput. Electron. Agric..

[8]  Meixin Zhu,et al.  Human-Like Autonomous Car-Following Model with Deep Reinforcement Learning , 2018, Transportation Research Part C: Emerging Technologies.

[9]  Marco Wiering,et al.  Traffic signal optimization through discrete and continuous reinforcement learning with robustness analysis in downtown Tehran , 2018, Adv. Eng. Informatics.

[10]  Lili Du,et al.  Cooperative platoon control for a mixed traffic flow including human drive vehicles and connected and autonomous vehicles , 2018, Transportation Research Part B: Methodological.

[11]  Yinhai Wang,et al.  Multistep speed prediction on traffic networks: A deep learning approach considering spatio-temporal dependencies , 2019, Transportation Research Part C: Emerging Technologies.

[12]  Zhongsheng Hou,et al.  Data-driven approximate Q-learning stabilization with optimality error bound analysis , 2019, Autom..

[13]  Lily Elefteriadou,et al.  Signal control optimization for automated vehicles at isolated signalized intersections , 2014 .

[14]  Alexandre M. Bayen,et al.  Expert Level Control of Ramp Metering Based on Multi-Task Deep Reinforcement Learning , 2017, IEEE Transactions on Intelligent Transportation Systems.

[15]  Etienne Perot,et al.  Deep Reinforcement Learning framework for Autonomous Driving , 2017, Autonomous Vehicles and Machines.

[16]  Xiao-Feng Xie,et al.  SIV-DSS: Smart In-Vehicle Decision Support System for driving at signalized intersections with V2I communication , 2018 .

[17]  Qing Li,et al.  Application and Evaluation of the Reinforcement Learning Approach to Eco-Driving at Intersections under Infrastructure-to-Vehicle Communications , 2018, Transportation Research Record: Journal of the Transportation Research Board.

[18]  Yiheng Feng,et al.  Integrated optimization of traffic signals and vehicle trajectories at isolated urban intersections , 2018, Transportation Research Part B: Methodological.