iPlanner: Imperative Path Planning

The problem of path planning has been studied for years. Classic planning pipelines, including perception, mapping, and path searching, can result in latency and compounding errors between modules. While recent studies have demonstrated the effectiveness of end-to-end learning methods in achieving high planning efficiency, these methods often struggle to match the generalization abilities of classic approaches in handling different environments. Moreover, end-to-end training of policies often requires a large number of labeled data or training iterations to reach convergence. In this paper, we present a novel Imperative Learning (IL) approach. This approach leverages a differentiable cost map to provide implicit supervision during policy training, eliminating the need for demonstrations or labeled trajectories. Furthermore, the policy training adopts a Bi-Level Optimization (BLO) process, which combines network update and metric-based trajectory optimization, to generate a smooth and collision-free path toward the goal based on a single depth measurement. The proposed method allows task-level costs of predicted trajectories to be backpropagated through all components to update the network through direct gradient descent. In our experiments, the method demonstrates around 4x faster planning than the classic approach and robustness against localization noise. Additionally, the IL approach enables the planner to generalize to various unseen environments, resulting in an overall 26-87% improvement in SPL performance compared to baseline learning methods.

[1]  S. Levine,et al.  GNM: A General Navigation Model to Drive Any Robot , 2022, ArXiv.

[2]  Brady G. Moon,et al.  PyPose: A Library for Robot Learning with Physics-based Optimization , 2022, 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[3]  Stephen Tu,et al.  Learning Model Predictive Controllers with Real-Time Attention for Real-World Navigation , 2022, CoRL.

[4]  Xianke Lin,et al.  High-Definition Map Generation Technologies For Autonomous Driving , 2022, ArXiv.

[5]  Jemin Hwangbo,et al.  Learning Forward Dynamics Model and Informed Trajectory Sampler for Safe Quadruped Navigation , 2022, Robotics: Science and Systems XVIII.

[6]  Howie Choset,et al.  Autonomous Exploration Development Environment and the Planning Algorithms , 2021, 2022 International Conference on Robotics and Automation (ICRA).

[7]  Deyu Meng,et al.  Investigating Bi-Level Optimization for Learning and Vision From a Unified Perspective: A Survey and Beyond , 2021, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[8]  René Ranftl,et al.  Learning high-speed flight in the wild , 2021, Science Robotics.

[9]  Dhruv Batra,et al.  Auxiliary Tasks and Exploration Enable ObjectGoal Navigation , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[10]  Lorenz Wellhausen,et al.  Rough Terrain Navigation for Legged Robots using Reachability Planning and Template Learning , 2021, 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[11]  Lorenz Wellhausen,et al.  Real-time Optimal Navigation Planning Using Learned Motion Costs , 2021, 2021 IEEE International Conference on Robotics and Automation (ICRA).

[12]  Lorenz Wellhausen,et al.  Learning a State Representation and Navigation in Cluttered and Dynamic Environments , 2021, IEEE Robotics and Automation Letters.

[13]  S. Levine,et al.  BADGR: An Autonomous Self-Supervised Learning-Based Navigation System , 2020, IEEE Robotics and Automation Letters.

[14]  Kostas Alexis,et al.  Complementary Multi–Modal Sensor Fusion for Resilient Robot Pose Estimation in Subterranean Environments , 2020, 2020 International Conference on Unmanned Aircraft Systems (ICUAS).

[15]  Chen Hu,et al.  Falco: Fast likelihood‐based collision avoidance with extension to human‐guided navigation , 2020, J. Field Robotics.

[16]  Kao-Shing Hwang,et al.  End-to-End Navigation Strategy With Deep Reinforcement Learning for Mobile Robots , 2020, IEEE Transactions on Industrial Informatics.

[17]  Xiaojun Chang,et al.  Vision-Language Navigation With Self-Supervised Auxiliary Reasoning Tasks , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Ari S. Morcos,et al.  DD-PPO: Learning Near-Perfect PointGoal Navigators from 2.5 Billion Frames , 2019, ICLR.

[19]  Jitendra Malik,et al.  On Evaluation of Embodied Navigation Agents , 2018, ArXiv.

[20]  Andreas Krause,et al.  Reinforced Imitation: Sample Efficient Deep Reinforcement Learning for Mapless Navigation by Leveraging Prior Demonstrations , 2018, IEEE Robotics and Automation Letters.

[21]  Matthias Nießner,et al.  Matterport3D: Learning from RGB-D Data in Indoor Environments , 2017, 2017 International Conference on 3D Vision (3DV).

[22]  Roland Siegwart,et al.  From perception to decision: A data-driven approach to end-to-end motion planning for autonomous ground robots , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[23]  Peter Fankhauser,et al.  ANYmal - a highly mobile and dynamic quadrupedal robot , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[24]  Xin Zhang,et al.  End to End Learning for Self-Driving Cars , 2016, ArXiv.

[25]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[26]  김정민,et al.  Cubic Spline Interpolation을 이용한 얼굴 영상의 단순화 , 2010 .

[27]  Andrew Howard,et al.  Design and use paradigms for Gazebo, an open-source multi-robot simulator , 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566).

[28]  Yoshua Bengio,et al.  Convolutional networks for images, speech, and time series , 1998 .

[29]  Michael A. Lachance,et al.  An introduction to splines for use in computer graphics and geometric modeling , 1990 .