Learning Interaction-aware Guidance Policies for Motion Planning in Dense Traffic Scenarios

Autonomous navigation in dense traffic scenarios remains challenging for autonomous vehicles (AVs) because the intentions of other drivers are not directly observable and AVs have to deal with a wide range of driving behaviors. To maneuver through dense traffic, AVs must be able to reason how their actions affect others (interaction model) and exploit this reasoning to navigate through dense traffic safely. This paper presents a novel framework for interaction-aware motion planning in dense traffic scenarios. We explore the connection between human driving behavior and their velocity changes when interacting. Hence, we propose to learn, via deep Reinforcement Learning (RL), an interaction-aware policy providing global guidance about the cooperativeness of other vehicles to an optimization-based planner ensuring safety and kinematic feasibility through constraint satisfaction. The learned policy can reason and guide the local optimization-based planner with interactive behavior to pro-actively merge in dense traffic while remaining safe in case the other vehicles do not yield. We present qualitative and quantitative results in highly interactive simulation environments (highway merging and unprotected left turns) against two baseline approaches, a learning-based and an optimization-based method. The presented results demonstrate that our method significantly reduces the number of collisions and increases the success rate with respect to both learning-based and optimization-based baselines.

[1]  Alexander Domahidi,et al.  FORCES NLP: an efficient implementation of interior-point methods for multistage nonlinear nonconvex programs , 2020, Int. J. Control.

[2]  Christoph Stiller,et al.  A Belief State Planner for Interactive Merge Maneuvers in Congested Traffic , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[3]  Sergey Levine,et al.  Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.

[4]  Javier Alonso-Mora,et al.  Joint Multi-Policy Behavior Estimation and Receding-Horizon Trajectory Planning for Automated Urban Driving , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[5]  Jonas Sjöberg,et al.  Learning Negotiating Behavior Between Cars in Intersections using Deep Q-Learning , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[6]  Marco Pavone,et al.  Multimodal Probabilistic Model-Based Planning for Human-Robot Interaction , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[7]  Omar Y. Al-Jarrah,et al.  Deep Learning-based Vehicle Behaviour Prediction For Autonomous Driving Applications: A Review , 2019, ArXiv.

[8]  Sergio Casas,et al.  Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations , 2020, ECCV.

[9]  Xin Zhang,et al.  End to End Learning for Self-Driving Cars , 2016, ArXiv.

[10]  Ching-Yao Chan,et al.  A Reinforcement Learning Based Approach for Automated Lane Change Maneuvers , 2018, 2018 IEEE Intelligent Vehicles Symposium (IV).

[11]  Emilio Frazzoli,et al.  A Survey of Motion Planning and Control Techniques for Self-Driving Urban Vehicles , 2016, IEEE Transactions on Intelligent Vehicles.

[12]  Lukas Hewing,et al.  Learning-Based Model Predictive Control: Toward Safe Learning in Control , 2020, Annu. Rev. Control. Robotics Auton. Syst..

[13]  Marcelo H. Ang,et al.  Situation-aware decision making for autonomous driving on urban road using online POMDP , 2015, 2015 IEEE Intelligent Vehicles Symposium (IV).

[14]  Jorge Pena Queralta,et al.  Sim-to-Real Transfer in Deep Reinforcement Learning for Robotics: a Survey , 2020, 2020 IEEE Symposium Series on Computational Intelligence (SSCI).

[15]  Helbing,et al.  Congested traffic states in empirical observations and microscopic simulations , 2000, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[16]  Gary A. Davis,et al.  Field Study of Gap Acceptance by Left-Turning Drivers , 2004 .

[17]  Dimitar Filev,et al.  Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning , 2019, Robotics Auton. Syst..

[18]  Javier Alonso-Mora,et al.  Curvature Aware Motion Planning with Closed-Loop Rapidly-exploring Random Trees , 2021, 2021 IEEE Intelligent Vehicles Symposium (IV).

[19]  Markus Maurer,et al.  Structuring Cooperative Behavior Planning Implementations for Automated Driving , 2015, 2015 IEEE 18th International Conference on Intelligent Transportation Systems.

[20]  Alexandre M. Bayen,et al.  Flow: Architecture and Benchmarking for Reinforcement Learning in Traffic Control , 2017, ArXiv.

[21]  Pete Trautman,et al.  Sparse interacting Gaussian processes: Efficiency and optimality theorems of autonomous crowd navigation , 2017, 2017 IEEE 56th Annual Conference on Decision and Control (CDC).

[22]  Dhruv Saxena,et al.  Cooperation-Aware Lane Change Maneuver in Dense Traffic based on Model Predictive Control with Recurrent Neural Network , 2019, 2020 American Control Conference (ACC).

[23]  Jonas Sjöberg,et al.  Learning When to Drive in Intersections by Combining Reinforcement Learning and Model Predictive Control , 2019, 2019 IEEE Intelligent Transportation Systems Conference (ITSC).

[24]  Christoph Stiller,et al.  Decision making for autonomous driving considering interaction and uncertain prediction of surrounding vehicles , 2017, 2017 IEEE Intelligent Vehicles Symposium (IV).

[25]  Mykel J. Kochenderfer,et al.  Cooperation-Aware Reinforcement Learning for Merging in Dense Traffic , 2019, 2019 IEEE Intelligent Transportation Systems Conference (ITSC).

[26]  Mykel J. Kochenderfer,et al.  Imitating driver behavior with generative adversarial networks , 2017, 2017 IEEE Intelligent Vehicles Symposium (IV).

[27]  Guy Rosman,et al.  Variational Autoencoder for End-to-End Control of Autonomous Driving with Novelty Detection and Training De-biasing , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[28]  Alexey Dosovitskiy,et al.  End-to-End Driving Via Conditional Imitation Learning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[29]  David González,et al.  A Review of Motion Planning Techniques for Automated Vehicles , 2016, IEEE Transactions on Intelligent Transportation Systems.

[30]  A. Zgonnikov,et al.  Should I stay or should I go? Evidence accumulation drives decision making in human drivers , 2020 .

[31]  J. Burdick,et al.  Safe Multi-Agent Interaction through Robust Control Barrier Functions with Learned Uncertainties , 2020, 2020 59th IEEE Conference on Decision and Control (CDC).

[32]  Mac Schwager,et al.  LUCIDGames: Online Unscented Inverse Dynamic Games for Adaptive Trajectory Prediction and Planning , 2020, IEEE Robotics and Automation Letters.

[33]  Anne Spalanzani,et al.  Game theoretic decision making for autonomous vehicles’ merge manoeuvre in high traffic scenarios , 2019, 2019 IEEE Intelligent Transportation Systems Conference (ITSC).

[34]  Mayank Bansal,et al.  ChauffeurNet: Learning to Drive by Imitating the Best and Synthesizing the Worst , 2018, Robotics: Science and Systems.

[35]  William Whittaker,et al.  Autonomous driving in urban environments: Boss and the Urban Challenge , 2008, J. Field Robotics.

[36]  Anca D. Dragan,et al.  Hierarchical Game-Theoretic Planning for Autonomous Vehicles , 2018, 2019 International Conference on Robotics and Automation (ICRA).

[37]  Kikuo Fujimura,et al.  Tactical Decision Making for Lane Changing with Deep Reinforcement Learning , 2017 .

[38]  David Hsu,et al.  Intention-aware online POMDP planning for autonomous driving in a crowd , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[39]  Haris N. Koutsopoulos,et al.  Modeling Integrated Lane-Changing Behavior , 2003 .

[40]  Dariu M. Gavrila,et al.  Human motion trajectory prediction: a survey , 2019, Int. J. Robotics Res..

[41]  Sebastian Thrun,et al.  Junior: The Stanford entry in the Urban Challenge , 2008, J. Field Robotics.

[42]  Andreas Krause,et al.  Unfreezing the robot: Navigation in dense, interacting crowds , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[43]  Yang Yu,et al.  Towards Sample Efficient Reinforcement Learning , 2018, IJCAI.

[44]  John M. Dolan,et al.  Traffic interaction in the urban challenge: Putting boss on its best behavior , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[45]  Colin Camerer,et al.  A Cognitive Hierarchy Model of Games , 2004 .

[46]  Dariu Gavrila,et al.  SafeVRU: A Research Platform for the Interaction of Self-Driving Vehicles with Vulnerable Road Users , 2019, 2019 IEEE Intelligent Vehicles Symposium (IV).

[47]  Mykel J. Kochenderfer,et al.  Reinforcement Learning with Probabilistic Guarantees for Autonomous Driving , 2019, ArXiv.

[48]  Javier Alonso-Mora,et al.  Safe Nonlinear Trajectory Generation for Parallel Autonomy With a Dynamic Vehicle Model , 2018, IEEE Transactions on Intelligent Transportation Systems.

[49]  Marco Pavone,et al.  MATS: An Interpretable Trajectory Forecasting Representation for Planning and Control , 2020, CoRL.

[50]  Chonhyon Park,et al.  I-Planner: Intention-aware motion planning using learning-based human motion prediction , 2016, Int. J. Robotics Res..

[51]  Jaime F. Fisac,et al.  Bridging Hamilton-Jacobi Safety Analysis and Reinforcement Learning , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[52]  Javier Alonso-Mora,et al.  Model Predictive Contouring Control for Collision Avoidance in Unstructured Dynamic Environments , 2019, IEEE Robotics and Automation Letters.

[53]  Nathan Fulton,et al.  Safe Reinforcement Learning via Formal Methods: Toward Safe Control Through Proof and Learning , 2018, AAAI.

[54]  Javier Alonso-Mora,et al.  Planning and Decision-Making for Autonomous Vehicles , 2018, Annu. Rev. Control. Robotics Auton. Syst..

[55]  Malte Risto,et al.  The social behavior of autonomous vehicles , 2016, UbiComp Adjunct.

[56]  Maxim Likhachev,et al.  Driving in Dense Traffic with Model-Free Reinforcement Learning , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[57]  David Isele,et al.  Reinforcement Learning with Iterative Reasoning for Merging in Dense Traffic , 2020, 2020 IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC).

[58]  Anca D. Dragan,et al.  Planning for Autonomous Cars that Leverage Effects on Human Actions , 2016, Robotics: Science and Systems.

[59]  Winnie Daamen,et al.  Key Variables of Merging Behaviour: Empirical Comparison between Two Sites and Assessment of Gap Acceptance Theory , 2013 .