Learning Autonomous Driving Styles and Maneuvers from Expert Demonstration

One of the many challenges in building robust and reliable autonomous systems is the large number of parameters and settings such systems often entail. The traditional approach to this task is simply to have system experts hand tune various parameter settings, and then validate them through simulation, offline playback, and field testing. However, this approach is tedious and time consuming for the expert, and typically produces subpar performance that does not generalize. Machine learning offers a solution to this problem in the form of learning from demonstration. Rather than ask an expert to explicitly encode his own preferences, he must simply demonstrate them, allowing the system to autonomously configure itself accordingly. This work extends this approach to the task of learning driving styles and maneuver preferences for an autonomous vehicle. Head to head experiments in simulation and with a live autonomous system demonstrate that this approach produces better autonomous performance, and with less expert interaction, than traditional hand tuning.

[1]  Stefan Schaal,et al.  Learning locomotion over rough terrain using terrain templates , 2009, 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[2]  David Silver,et al.  Learning to search: Functional gradient techniques for imitation learning , 2009, Auton. Robots.

[3]  Alonzo Kelly,et al.  Toward Reliable Off Road Autonomous Vehicles Operating in Challenging Environments , 2006, Int. J. Robotics Res..

[4]  Brett Browning,et al.  A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..

[5]  Sebastian Thrun,et al.  Junior: The Stanford entry in the Urban Challenge , 2008, J. Field Robotics.

[6]  William Whittaker,et al.  Autonomous driving in urban environments: Boss and the Urban Challenge , 2008, J. Field Robotics.

[7]  Pieter Abbeel,et al.  Hierarchical Apprenticeship Learning with Application to Quadruped Locomotion , 2007, NIPS.

[8]  Pieter Abbeel,et al.  Autonomous Helicopter Aerobatics through Apprenticeship Learning , 2010, Int. J. Robotics Res..

[9]  David Silver,et al.  Active learning from demonstration for robust autonomous navigation , 2012, 2012 IEEE International Conference on Robotics and Automation.

[10]  William Whittaker,et al.  Autonomous driving in urban environments: Boss and the Urban Challenge , 2008, J. Field Robotics.

[11]  Sebastian Scherer,et al.  Learning obstacle avoidance parameters from operator behavior , 2006, J. Field Robotics.

[12]  Heiko Hirschmüller,et al.  Stereo-vision-based navigation of a six-legged walking robot in unknown rough terrain , 2012, Int. J. Robotics Res..

[13]  Christopher G. Atkeson,et al.  Optimization and learning for rough terrain legged locomotion , 2011, Int. J. Robotics Res..

[14]  James M. Rehg,et al.  Learning from examples in unstructured, outdoor environments , 2006, J. Field Robotics.

[15]  Peter L. Bartlett,et al.  Boosting Algorithms as Gradient Descent in Function Space , 2007 .

[16]  Sebastian Thrun,et al.  Apprenticeship learning for motion planning with application to parking lot navigation , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[17]  Reid G. Simmons,et al.  Recent progress in local and global traversability for planetary rovers , 2000, Proceedings 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation. Symposia Proceedings (Cat. No.00CH37065).

[18]  David Silver,et al.  Learning Preference Models for Autonomous Mobile Robots in Complex Domains , 2010 .

[19]  David Silver,et al.  Learning from Demonstration for Autonomous Navigation in Complex Unstructured Terrain , 2010, Int. J. Robotics Res..

[20]  David M. Bradley,et al.  Learning for Autonomous Navigation , 2010, IEEE Robotics & Automation Magazine.

[21]  Peter L. Bartlett,et al.  Boosting Algorithms as Gradient Descent , 1999, NIPS.

[22]  Peter King,et al.  Odin: Team VictorTango's entry in the DARPA Urban Challenge , 2008, J. Field Robotics.