Adversarial Evaluation of Autonomous Vehicles in Lane-Change Scenarios

Autonomous vehicles must be comprehensively evaluated before deployed in cities and highways. Current evaluation procedures lack the abilities of weakness-aiming and evolving, thus they could hardly generate adversarial environments for autonomous vehicles, leading to insufficient challenges. To overcome the shortage of static evaluation methods, this paper proposes a novel method to generate adversarial environments with deep reinforcement learning, and to cluster them with a nonparametric Bayesian method. As a representative task of autonomous driving, lane-change is used to demonstrate the superiority of the proposed method. First, two lane-change models are separately developed by a rule-based method and a learning-based method, waiting for evaluation and comparison. Next, adversarial environments are generated by training surrounding interactive vehicles with deep reinforcement learning for local optimal ensembles. Then, a nonparametric Bayesian approach is utilized to cluster the adversarial policies of the interactive vehicles. Finally, the adversarial environment patterns are illustrated and the performances of two lane-change models are evaluated and compared. The simulation results indicate that both models perform significantly worse in adversarial environments than in naturalistic environments, with plenty of weaknesses successfully extracted in a few tests.

[1]  Paul Geladi,et al.  Principal Component Analysis , 1987, Comprehensive Chemometrics.

[2]  Manfred Morari,et al.  Model predictive control: Theory and practice - A survey , 1989, Autom..

[3]  Leemon C. Baird,et al.  Residual Algorithms: Reinforcement Learning with Function Approximation , 1995, ICML.

[4]  Moshe Ben-Akiva,et al.  MODELS OF FREEWAY LANE CHANGING AND GAP ACCEPTANCE BEHAVIOR , 1996 .

[5]  Yishay Mansour,et al.  Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.

[6]  Helbing,et al.  Congested traffic states in empirical observations and microscopic simulations , 2000, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[7]  Haris N. Koutsopoulos,et al.  Modeling Integrated Lane-Changing Behavior , 2003 .

[8]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[9]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[10]  Anirban Mahanti,et al.  Traffic classification using clustering algorithms , 2006, MineNet '06.

[11]  Dot Hs,et al.  Analysis of Lane-Change Crashes and Near-Crashes , 2009 .

[12]  Julius Ziegler,et al.  Optimal trajectory generation for dynamic street scenarios in a Frenét Frame , 2010, 2010 IEEE International Conference on Robotics and Automation.

[13]  Michael I. Jordan,et al.  Revisiting k-means: New Algorithms via Bayesian Nonparametrics , 2011, ICML.

[14]  Guy Lever,et al.  Deterministic Policy Gradient Algorithms , 2014, ICML.

[15]  Andreas Krause,et al.  Coresets for Nonparametric Estimation - the Case of DP-Means , 2015, ICML.

[16]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[17]  Yuval Tassa,et al.  Continuous control with deep reinforcement learning , 2015, ICLR.

[18]  Stefano Ermon,et al.  Generative Adversarial Imitation Learning , 2016, NIPS.

[19]  Germán Ros,et al.  CARLA: An Open Urban Driving Simulator , 2017, CoRL.

[20]  Ding Zhao,et al.  Accelerated Evaluation of Automated Vehicles. , 2016 .

[21]  Jonas Fredriksson,et al.  Lane Change Maneuvers for Automated Vehicles , 2017, IEEE Transactions on Intelligent Transportation Systems.

[22]  Balaraman Ravindran,et al.  EPOpt: Learning Robust Neural Network Policies Using Model Ensembles , 2016, ICLR.

[23]  Abhinav Gupta,et al.  Robust Adversarial Reinforcement Learning , 2017, ICML.

[24]  Alec Radford,et al.  Proximal Policy Optimization Algorithms , 2017, ArXiv.

[25]  Shimon Whiteson,et al.  OFFER: Off-Environment Reinforcement Learning , 2017, AAAI.

[26]  Philip Koopman,et al.  Autonomous Vehicle Safety: An Interdisciplinary Challenge , 2017, IEEE Intelligent Transportation Systems Magazine.

[27]  M. Endsley Autonomous Driving Systems: A Preliminary Naturalistic Study of the Tesla Model S , 2017 .

[28]  John M. Dolan,et al.  Automatically Generated Curriculum based Reinforcement Learning for Autonomous Vehicles in Urban Environment , 2018, 2018 IEEE Intelligent Vehicles Symposium (IV).

[29]  Kenneth O. Stanley,et al.  Improving Exploration in Evolution Strategies for Deep Reinforcement Learning via a Population of Novelty-Seeking Agents , 2017, NeurIPS.

[30]  Felix Bießmann,et al.  Automating Large-Scale Data Quality Verification , 2018, Proc. VLDB Endow..

[31]  Jakub W. Pachocki,et al.  Emergent Complexity via Multi-Agent Competition , 2017, ICLR.

[32]  Bernard Ghanem,et al.  Driving Policy Transfer via Modularity and Abstraction , 2018, CoRL.

[33]  Kanika Chourasia,et al.  Autonomous vehicles: challenges, opportunities, and future implications for transportation policies , 2019 .

[34]  Ching-Yao Chan,et al.  Driving Decision and Control for Automated Lane Change Behavior based on Deep Reinforcement Learning , 2019, 2019 IEEE Intelligent Transportation Systems Conference (ITSC).

[35]  Francesco Braghin,et al.  Autonomous Vehicle Driving via Deep Deterministic Policy Gradient , 2019 .

[36]  Nanning Zheng,et al.  Parallel testing of vehicle intelligence via virtual-real interaction , 2019, Science Robotics.

[37]  Zhenbo Lu,et al.  Interval data‐based k ‐means clustering method for traffic state identification at urban intersections , 2019, IET Intelligent Transport Systems.

[38]  M. Shawky Factors affecting lane change crashes , 2020 .

[39]  Qi Sun,et al.  Hierarchical Reinforcement Learning for Self-Driving Decision-Making without Reliance on Labeled Driving Data , 2020, IET Intelligent Transport Systems.

[40]  Yi Zhang,et al.  Testing Scenario Library Generation for Connected and Automated Vehicles, Part I: Methodology , 2019, IEEE Transactions on Intelligent Transportation Systems.