Terrain-Aware Learned Controllers for Sampling-Based Kinodynamic Planning over Physically Simulated Terrains

This paper explores learning an effective controller for improving the efficiency of kinodynamic planning for vehicular systems navigating uneven terrains. It describes the pipeline for training the corresponding controller and using it for motion planning purposes. The training process uses a soft actor-critic approach with hindsight experience replay to train a model, which is parameterized by the incline of the robot's local terrain. This trained model is then used during the expansion process of an asymptotically optimal kinodynamic planner to generate controls that allow the robot to reach desired local states. It is also used to define a heuristic cost-to-go function for the planner via a wavefront operation that estimates the cost of reaching the global goal. The cost-to-go function is used both for selecting nodes for expansion as well as for generating local goals for the controller to expand towards. The accompanying experimental section applies the integrated planning solution on models of all-terrain robots in a variety of physically simulated terrains. It shows that the proposed terrain-aware controller and the proposed wavefront function based on the cost-to-go model enable motion planners to find solutions in less time and with lower cost than alternatives. An ablation study emphasizes the benefits of a learned controller that is parameterized by the incline of the robot's local terrain as well as of an incremental training process for the controller.

[1]  Troy McMahon,et al.  Improving Kinodynamic Planners for Vehicular Navigation with Learned Goal-Reaching Controllers , 2021, 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[2]  P. Stone,et al.  Learning Inverse Kinodynamics for Accurate High-Speed Off-Road Navigation on Unstructured Terrain , 2021, IEEE Robotics and Automation Letters.

[3]  S. Levine,et al.  BADGR: An Autonomous Self-Supervised Learning-Based Navigation System , 2020, IEEE Robotics and Automation Letters.

[4]  Michael C. Yip,et al.  Motion Planning Networks: Bridging the Gap Between Learning-Based and Classical Motion Planners , 2019, IEEE Transactions on Robotics.

[5]  Michael C. Yip,et al.  Dynamically Constrained Motion Planning Networks for Non-Holonomic Robots , 2020, 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[6]  Jiankun Wang,et al.  Neural RRT*: Learning-Based Optimal Path Planning , 2020, IEEE Transactions on Automation Science and Engineering.

[7]  Edward Schmerling,et al.  Learned Critical Probabilistic Roadmaps for Robotic Motion Planning , 2019 .

[8]  Joonho Lee,et al.  DeepGait: Planning and Control of Quadrupedal Gaits Using Deep Reinforcement Learning , 2019, IEEE Robotics and Automation Letters.

[9]  Kostas E. Bekris,et al.  Refined Analysis of Asymptotically-Optimal Kinodynamic Planning in the State-Cost Space , 2019, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[10]  Natalia Gimelshein,et al.  PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.

[11]  Siddhartha S. Srinivasa,et al.  LEGO: Leveraging Experience in Roadmap Generation for Sampling-Based Planning , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[12]  Lydia Tapia,et al.  RL-RRT: Kinodynamic Motion Planning via Learning Reachability Estimators From RL Policies , 2019, IEEE Robotics and Automation Letters.

[13]  Aviv Tamar,et al.  Harnessing Reinforcement Learning for Neural Motion Planning , 2019, Robotics: Science and Systems.

[14]  George P. Kontoudis,et al.  Kinodynamic Motion Planning With Continuous-Time Q-Learning: An Online, Model-Free, and Safe Navigation Framework , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[15]  Kostas E. Bekris,et al.  Efficient and Asymptotically Optimal Kinodynamic Motion Planning via Dominance-Informed Regions , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[16]  Zhijun Li,et al.  Neural Network Approximation Based Near-Optimal Motion Planning With Kinodynamic Constraints Using RRT , 2018, IEEE Transactions on Industrial Electronics.

[17]  Sergey Levine,et al.  Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.

[18]  Lydia Tapia,et al.  PRM-RL: Long-range Robotic Navigation Tasks by Combining Reinforcement Learning and Sampling-Based Planning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[19]  Marco Pavone,et al.  Learning Sampling Distributions for Robot Motion Planning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[20]  Marcin Andrychowicz,et al.  Hindsight Experience Replay , 2017, NIPS.

[21]  Wojciech Zaremba,et al.  OpenAI Gym , 2016, ArXiv.

[22]  Didier Devaurs,et al.  Optimal Path Planning in Complex Cost Spaces With Sampling-Based Algorithms , 2016, IEEE Transactions on Automation Science and Engineering.

[23]  Kris Hauser,et al.  Asymptotically Optimal Planning by Feasible Kinodynamic Planning in a State–Cost Space , 2015, IEEE Transactions on Robotics.

[24]  Kostas E. Bekris,et al.  Asymptotically optimal sampling-based kinodynamic planning , 2014, Int. J. Robotics Res..

[25]  Jur P. van den Berg,et al.  Kinodynamic RRT*: Asymptotically optimal motion planning for robots with linear dynamics , 2013, 2013 IEEE International Conference on Robotics and Automation.

[26]  Didier Devaurs,et al.  Enhancing the transition-based RRT to deal with complex cost spaces , 2013, 2013 IEEE International Conference on Robotics and Automation.

[27]  Emilio Frazzoli,et al.  Sampling-based algorithms for optimal motion planning , 2011, Int. J. Robotics Res..

[28]  David Silver,et al.  Learning from Demonstration for Autonomous Navigation in Complex Unstructured Terrain , 2010, Int. J. Robotics Res..

[29]  Jason Weston,et al.  Curriculum learning , 2009, ICML '09.

[30]  Thierry Siméon,et al.  Transition-based RRT for path planning in continuous cost spaces , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[31]  Steven M. LaValle,et al.  Randomized Kinodynamic Planning , 1999, Proceedings 1999 IEEE International Conference on Robotics and Automation (Cat. No.99CH36288C).

[32]  Rajeev Motwani,et al.  Path planning in expansive configuration spaces , 1997, Proceedings of International Conference on Robotics and Automation.

[33]  B. Faverjon,et al.  Probabilistic Roadmaps for Path Planning in High-Dimensional Con(cid:12)guration Spaces , 1996 .