Human-like planning of swerve maneuvers for autonomous vehicles

In this paper, we develop a motion planner for on-road autonomous swerve maneuvers that is capable of learning passengers' individual driving styles. It uses a hybrid planning approach that combines sampling-based graph search and vehicle model-based evaluation to obtain a smooth trajectory plan. To automate the parameter tuning process, as well as to reflect individual driving styles, we further adapt inverse reinforcement learning techniques to distill human driving patterns from maneuver demonstrations collected from different individuals. We found that the proposed swerve planner and its learning routine can approximate a good variety of maneuver demonstrations. However, due to the underlying stochastic nature of human driving, more data are needed in order to obtain a more generative swerve model.

[1]  David Silver,et al.  Learning from Demonstration for Autonomous Navigation in Complex Unstructured Terrain , 2010, Int. J. Robotics Res..

[2]  Jin-Woo Lee,et al.  Motion planning for autonomous driving with a conformal spatiotemporal lattice , 2011, 2011 IEEE International Conference on Robotics and Automation.

[3]  Julius Ziegler,et al.  Trajectory planning for Bertha — A local, continuous method , 2014, 2014 IEEE Intelligent Vehicles Symposium Proceedings.

[4]  Jing Wang,et al.  A suboptimal and analytical solution to mobile robot trajectory generation amidst moving obstacles , 2015, Auton. Robots.

[5]  Sebastian Thrun,et al.  Stanley: The robot that won the DARPA Grand Challenge , 2006, J. Field Robotics.

[6]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[7]  Dinh Quoc Tran,et al.  An application of sequential convex programming to time optimal trajectory planning for a car motion , 2009, Proceedings of the 48h IEEE Conference on Decision and Control (CDC) held jointly with 2009 28th Chinese Control Conference.

[8]  Qi Zhu,et al.  A sampling-based local trajectory planner for autonomous driving along a reference path , 2014, 2014 IEEE Intelligent Vehicles Symposium Proceedings.

[9]  Sebastian Scherer,et al.  Learning obstacle avoidance parameters from operator behavior , 2006, J. Field Robotics.

[10]  Pieter Abbeel,et al.  Finding Locally Optimal, Collision-Free Trajectories with Sequential Convex Optimization , 2013, Robotics: Science and Systems.

[11]  Pieter Abbeel,et al.  Apprenticeship learning and reinforcement learning with application to robotic control , 2008 .

[12]  Jin-Woo Lee,et al.  A unified framework of the automated lane centering/changing control for motion smoothness adaptation , 2012, 2012 15th International IEEE Conference on Intelligent Transportation Systems.

[13]  Julius Ziegler,et al.  Spatiotemporal state lattices for fast trajectory planning in dynamic on-road driving scenarios , 2009, 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[14]  D. Dolgov Practical Search Techniques in Path Planning for Autonomous Driving , 2008 .

[15]  Pierre Gançarski,et al.  A global averaging method for dynamic time warping, with applications to clustering , 2011, Pattern Recognit..

[16]  John M. Dolan,et al.  Toward human-like motion planning in urban environments , 2014, 2014 IEEE Intelligent Vehicles Symposium Proceedings.

[17]  William Whittaker,et al.  Autonomous driving in urban environments: Boss and the Urban Challenge , 2008, J. Field Robotics.