The Reasonable Crowd: Towards evidence-based and interpretable models of driving behavior

Autonomous vehicles must balance a complex set of objectives. There is no consensus on how they should do so, nor on a model for specifying a desired driving behavior. We created a dataset to help address some of these questions in a limited operating domain. The data consists of 92 traffic scenarios, with multiple ways of traversing each scenario. Multiple annotators expressed their preference between pairs of scenario traversals. We used the data to compare an instance of a rulebook [1], carefully hand-crafted independently of the dataset, with several interpretable machine learning models such as Bayesian networks, decision trees, and logistic regression trained on the dataset. To compare driving behavior, these models use scores indicating by how much different scenario traversals violate each of 14 driving rules. The rules are interpretable and designed by subject-matter experts. First, we found that these rules were enough for these models to achieve a high classification accuracy on the dataset. Second, we found that the rulebook provides high interpretability without excessively sacrificing performance. Third, the data pointed to possible improvements in the rulebook and the rules, and to potential new rules. Fourth, we explored the interpretability vs performance trade-off by also training non-interpretable models such as a random forest. Finally, we make the dataset publicly available to encourage a discussion from the wider community on behavior specification for AVs. Please find it at github.com/bassam-motional/Reasonable-Crowd.

[1]  Eric M. Wolff,et al.  nuPlan: A closed-loop ML-based planning benchmark for autonomous vehicles , 2021, ArXiv.

[2]  Calin Belta,et al.  Rule-based optimal control for autonomous driving , 2021, ICCPS.

[3]  Rules of the Road: Towards Safety and Liveness Guarantees for Autonomous Vehicles , 2020, ArXiv.

[4]  Scott Pendleton,et al.  Safety of the Intended Driving Behavior Using Rulebooks , 2020, 2020 IEEE Intelligent Vehicles Symposium (IV).

[5]  Mykel J. Kochenderfer,et al.  Interpretable Safety Validation for Autonomous Vehicles , 2020, 2020 IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC).

[6]  Qiang Xu,et al.  nuScenes: A Multimodal Dataset for Autonomous Driving , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Timothy Ha,et al.  Deep Predictive Autonomous Driving Using Multi-Agent Joint Trajectory Prediction and Traffic Rules , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[8]  Ersin Yumer,et al.  Jointly Learnable Behavior and Trajectory Planning for Self-Driving Vehicles , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[9]  Richard M. Murray,et al.  Towards Assume-Guarantee Profiles for Autonomous Vehicles , 2019, 2019 IEEE 58th Conference on Decision and Control (CDC).

[10]  Lars Svensson,et al.  Pre-Crash Vehicle Control and Manoeuvre Planning: A Step Towards Minimizing Collision Severity for Highly Automated Vehicles , 2019, 2019 IEEE International Conference of Vehicular Electronics and Safety (ICVES).

[11]  Dorsa Sadigh,et al.  Learning Reward Functions by Integrating Human Demonstrations and Preferences , 2019, Robotics: Science and Systems.

[12]  Sergio Casas,et al.  End-To-End Interpretable Neural Motion Planner , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[13]  Emilio Frazzoli,et al.  Liability, Ethics, and Culture-Aware Behavior Specification using Rulebooks , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[14]  Shane Legg,et al.  Reward learning from human preferences and demonstrations in Atari , 2018, NeurIPS.

[15]  Sebastian Raschka,et al.  Model Evaluation, Model Selection, and Algorithm Selection in Machine Learning , 2018, ArXiv.

[16]  J. Henrich,et al.  The Moral Machine experiment , 2018, Nature.

[17]  Dorsa Sadigh,et al.  Batch Active Preference-Based Learning of Reward Functions , 2018, CoRL.

[18]  Jungseock Joo,et al.  Protest Activity Detection and Perceived Violence Estimation from Social Media Images , 2017, ACM Multimedia.

[19]  Amnon Shashua,et al.  On a Formal Model of Safe and Scalable Self-driving Cars , 2017, ArXiv.

[20]  Markus Maurer,et al.  Towards a skill- and ability-based development process for self-aware automated road vehicles , 2017, 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC).

[21]  Anca D. Dragan,et al.  Active Preference-Based Learning of Reward Functions , 2017, Robotics: Science and Systems.

[22]  Sergio Escalera,et al.  ChaLearn looking at people: A review of events and resources , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).

[23]  Markus Wulfmeier,et al.  Watch this: Scalable cost-function learning for path planning in urban environments , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[24]  Sergey Levine,et al.  Guided Cost Learning: Deep Inverse Optimal Control via Policy Optimization , 2016, ICML.

[25]  Thorsten Joachims,et al.  Learning preferences for manipulation tasks from online coactive feedback , 2015, Int. J. Robotics Res..

[26]  Markus Wulfmeier,et al.  Maximum Entropy Deep Inverse Reinforcement Learning , 2015, 1507.04888.

[27]  Wolfram Burgard,et al.  Learning driving styles for autonomous vehicles from demonstration , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[28]  Jayesh K. Gupta,et al.  PlanIt: A crowdsourcing approach for learning to plan paths from large scale preference feedback , 2014, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[29]  Fabien Moutarde,et al.  Priority-based coordination of autonomous and legacy vehicles at intersection , 2014, 17th International IEEE Conference on Intelligent Transportation Systems (ITSC).

[30]  Markus Maurer,et al.  Probabilistic online POMDP decision making for lane changes in fully automated driving , 2013, 16th International IEEE Conference on Intelligent Transportation Systems (ITSC 2013).

[31]  Sergey Levine,et al.  Nonlinear Inverse Reinforcement Learning with Gaussian Processes , 2011, NIPS.

[32]  Sebastian Thrun,et al.  Apprenticeship learning for motion planning with application to parking lot navigation , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[33]  Nir Friedman,et al.  Bayesian Network Classifiers , 1997, Machine Learning.

[34]  R. A. Bradley,et al.  RANK ANALYSIS OF INCOMPLETE BLOCK DESIGNS THE METHOD OF PAIRED COMPARISONS , 1952 .

[35]  F. Galton Vox Populi , 1907, Nature.