Proactive Car-Following Using Deep-Reinforcement Learning

Car-following is a fundamental operation for vehicle control for both ADAS on modern vehicles and vehilce control on autonomous vehicles. Most existing car following mechanisms react to the observations of nearby vehicles in real-time. Unfortunately, lack of capability of taking into account multiple constraints and objectives, these mechanisms lead to poor efficiency, discomfort, and unsafe operations. In this paper, we design and implement a proactive car-following model to take into account safety regulation, efficiency, and comfort using deep reinforcement learning. The evaluation results show that the proactive model not only reduces the number of inefficient and unsafe headway but also eliminates the traffic jerk, compared to human drivers. The model outperformed 79% human drivers in public data set and the road efficiency is only 2% less than the optimal bound. Compared to ACC model, the DDPG model allows 4.4% more vehicles to finish the simulation than ACC model does, and increases the average speed for 28.4%.

[1]  Josef F. Krems,et al.  Comfort in automated driving: An analysis of preferences for different automated driving styles and their dependence on personality traits , 2018 .

[2]  Yuval Tassa,et al.  Continuous control with deep reinforcement learning , 2015, ICLR.

[3]  Lutz Eckstein,et al.  The highD Dataset: A Drone Dataset of Naturalistic Vehicle Trajectories on German Highways for Validation of Highly Automated Driving Systems , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[4]  Milan Simic,et al.  In the Passenger Seat: Investigating Ride Comfort Measures in Autonomous Cars , 2015, IEEE Intelligent Transportation Systems Magazine.

[5]  Antonio Lucas Alba,et al.  Car following: Comparing distance-oriented vs. inertia-oriented driving techniques , 2017, Transport Policy.

[6]  Feng Zhu,et al.  Modeling the Proactive Driving Behavior of Connected Vehicles: A Cell‐Based Simulation Approach , 2018, Comput. Aided Civ. Infrastructure Eng..

[7]  Michael Sivak,et al.  A Survey of Public Opinion about Autonomous and Self-Driving Vehicles in the U.S., the U.K., and Australia , 2014 .

[8]  Yun-Pang Flötteröd,et al.  Microscopic Traffic Simulation using SUMO , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[9]  Tarek Sayed,et al.  Surrogate Safety Assessment Model and Validation: Final Report , 2008 .

[10]  Meixin Zhu,et al.  Safe, Efficient, and Comfortable Velocity Control based on Reinforcement Learning for Autonomous Driving , 2019, ArXiv.

[11]  L. A. Pipes An Operational Analysis of Traffic Dynamics , 1953 .

[12]  Marco Pavone,et al.  Multi-objective optimal control for proactive decision making with temporal logic models , 2019, Int. J. Robotics Res..

[13]  Meng Wang,et al.  Realistic Car-Following Models for Microscopic Simulation of Adaptive and Cooperative Adaptive Cruise Control Vehicles , 2017 .

[14]  Roberto Zanasi,et al.  Head-neck model for the evaluation of passenger's comfort , 2002, IEEE/RSJ International Conference on Intelligent Robots and Systems.

[15]  Steven E Shladover,et al.  Modeling cooperative and autonomous adaptive cruise control dynamic responses using experimental data , 2014 .

[16]  Charles Desjardins,et al.  Cooperative Adaptive Cruise Control: A Reinforcement Learning Approach , 2011, IEEE Transactions on Intelligent Transportation Systems.

[17]  Ching-Yao Chan Advancements, prospects, and impacts of automated driving systems , 2017 .

[18]  Sen Wang,et al.  Deep Reinforcement Learning for Autonomous Driving , 2018, ArXiv.

[19]  Jean Elsner Optimizing Passenger Comfort in Cost Functions for Trajectory Planning , 2018, ArXiv.