Distance metric approximation for state-space RRTs using supervised learning

The dynamic feasibility of solutions to motion planning problems using Rapidly Exploring Random Trees depends strongly on the choice of the distance metric used while planning. The ideal distance metric is the optimal cost of traversal between two states in the state space. However, it is computationally intensive to find the optimal cost while planning. We propose a novel approach to overcome this barrier by using a supervised learning algorithm that learns a nonlinear function which is an estimate of the optimal cost, via offline training. We use the Iterative Linear Quadratic Regulator approach for estimating an approximation to the optimal cost and learn this cost using Locally Weighted Projection Regression. We show that the learnt function approximates the original cost with a reasonable tolerance and more importantly, gives a tremendous speed up of a factor of 1000 over the actual computation time. We also use the learnt metric for solving the pendulum swing up planning problem and show that our metric performs better than the popularly used Linear Quadratic Regulator based metric.

[1]  Steven M. LaValle,et al.  Improving Motion-Planning Algorithms by Efficient Nearest-Neighbor Searching , 2007, IEEE Transactions on Robotics.

[2]  Peter I. Corke,et al.  Robotics, Vision and Control - Fundamental Algorithms in MATLAB® , 2011, Springer Tracts in Advanced Robotics.

[3]  Steven M. LaValle,et al.  Randomized Kinodynamic Planning , 1999, Proceedings 1999 IEEE International Conference on Robotics and Automation (Cat. No.99CH36288C).

[4]  Emanuel Todorov,et al.  Iterative Linear Quadratic Regulator Design for Nonlinear Biological Movement Systems , 2004, ICINCO.

[5]  Leslie Pack Kaelbling,et al.  LQR-RRT*: Optimal sampling-based motion planning with automatically derived extension heuristics , 2012, 2012 IEEE International Conference on Robotics and Automation.

[6]  Matthew R. Walter,et al.  Reachability-guided sampling for planning under differential constraints , 2009, 2009 IEEE International Conference on Robotics and Automation.

[7]  Steven M. LaValle,et al.  Planning algorithms , 2006 .

[8]  Steven M. LaValle,et al.  From Dynamic Programming to RRTs: Algorithmic Design of Feasible Trajectories , 2003, Control Problems in Robotics.

[9]  Inna Sharf,et al.  Kinodynamic Motion Planning for UAVs: A Minimum Energy Approach , 2013 .

[10]  Jean-Claude Latombe,et al.  Randomized Kinodynamic Motion Planning with Moving Obstacles , 2002, Int. J. Robotics Res..

[11]  Stefan Schaal,et al.  Incremental Online Learning in High Dimensions , 2005, Neural Computation.

[12]  Russ Tedrake,et al.  A quadratic regulator-based heuristic for rapidly exploring state space , 2010, 2010 IEEE International Conference on Robotics and Automation.

[13]  Stefan Schaal,et al.  A Library for Locally Weighted Projection Regression , 2008, J. Mach. Learn. Res..