Bayesian Learning for Safe High-Speed Navigation in Unknown Environments

In this work, we develop a planner for high-speed navigation in unknown environments, for example reaching a goal in an unknown building in minimum time, or flying as fast as possible through a forest. This planning task is challenging because the distribution over possible maps, which is needed to estimate the feasibility and cost of trajectories, is unknown and extremely hard to model for real-world environments. At the same time, the worst-case assumptions that a receding-horizon planner might make about the unknown regions of the map may be overly conservative, and may limit performance. Therefore, robots must make accurate predictions about what will happen beyond the map frontiers to navigate as fast as possible. To reason about uncertainty in the map, we model this problem as a POMDP and discuss why it is so difficult given that we have no accurate probability distribution over real-world environments. We then present a novel method of predicting collision probabilities based on training data, which compensates for the missing environment distribution and provides an approximate solution to the POMDP. Extending our previous work, the principal result of this paper is that by using a Bayesian non-parametric learning algorithm that encodes formal safety constraints as a prior over collision probabilities, our planner seamlessly reverts to safe behavior when it encounters a novel environment for which it has no relevant training data. This strategy generalizes our method across all environment types, including those for which we have training data as well as those for which we do not. In familiar environment types with dense training data, we show an 80% speed improvement compared to a planner that is constrained to guarantee safety. In experiments, our planner has reached over 8 m/s in unknown cluttered indoor spaces. Video of our experimental demonstration is available at http://groups.csail.mit.edu/rrg/bayesian_learning_high_speed_nav.

[1]  Reid G. Simmons,et al.  The curvature-velocity method for local obstacle avoidance , 1996, Proceedings of IEEE International Conference on Robotics and Automation.

[2]  Brian Yamauchi,et al.  A frontier-based approach for autonomous exploration , 1997, Proceedings 1997 IEEE International Symposium on Computational Intelligence in Robotics and Automation CIRA'97. 'Towards New Computational Principles for Robotics and Automation'.

[3]  Wolfram Burgard,et al.  The dynamic window approach to collision avoidance , 1997, IEEE Robotics Autom. Mag..

[4]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[5]  Wolfram Burgard,et al.  Map learning and high-speed navigation in RHINO , 1998 .

[6]  Paolo Fiorini,et al.  Motion Planning in Dynamic Environments Using Velocity Obstacles , 1998, Int. J. Robotics Res..

[7]  Alexei Makarenko,et al.  An experiment in integrated exploration , 2002, IEEE/RSJ International Conference on Intelligent Robots and Systems.

[8]  Joelle Pineau,et al.  Point-based value iteration: An anytime algorithm for POMDPs , 2003, IJCAI.

[9]  Hajime Asama,et al.  Inevitable collision states. A step towards safer robots? , 2003, Proceedings 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453).

[10]  J. How,et al.  Receding horizon path planning with implicit safety guarantees , 2004, Proceedings of the 2004 American Control Conference.

[11]  Ashutosh Saxena,et al.  High speed obstacle avoidance using monocular vision and reinforcement learning , 2005, ICML.

[12]  Yann LeCun,et al.  Off-Road Obstacle Avoidance through End-to-End Learning , 2005, NIPS.

[13]  Wolfram Burgard,et al.  Information Gain-based Exploration Using Rao-Blackwellized Particle Filters , 2005, Robotics: Science and Systems.

[14]  Kostas E. Bekris,et al.  Greedy but Safe Replanning under Kinodynamic Constraints , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[15]  Thierry Fraichard,et al.  A Short Paper about Motion Safety , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[16]  David Hsu,et al.  SARSOP: Efficient Point-Based POMDP Planning by Approximating Optimally Reachable Belief Spaces , 2008, Robotics: Science and Systems.

[17]  Alonzo Kelly,et al.  State space sampling of feasible motions for high‐performance mobile robot navigation in complex environments , 2008, J. Field Robotics.

[18]  Urs A. Muller,et al.  Learning long-range vision for autonomous off-road driving , 2009 .

[19]  Leslie Pack Kaelbling,et al.  Collision Avoidance for Unmanned Aircraft using Markov Decision Processes , 2010 .

[20]  Joel Veness,et al.  Monte-Carlo Planning in Large POMDPs , 2010, NIPS.

[21]  Nicholas Roy,et al.  RANGE–Robust autonomous navigation in GPS‐denied environments , 2011, J. Field Robotics.

[22]  Nicholas Roy,et al.  Unmanned Aircraft Collision Avoidance Using Continuous-State POMDPs , 2012 .

[23]  Nicholas Roy,et al.  State estimation for aggressive flight in GPS-denied environments using onboard sensing , 2012, 2012 IEEE International Conference on Robotics and Automation.

[24]  Martin Buss,et al.  Safety assessment of robot trajectories for navigation in uncertain and dynamic environments , 2011, Autonomous Robots.

[25]  Martial Hebert,et al.  Learning monocular reactive UAV control in cluttered natural environments , 2012, 2013 IEEE International Conference on Robotics and Automation.

[26]  Ross A. Knepper,et al.  Model-Predictive Motion Planning: Several Key Developments for Autonomous Mobile Robots , 2014, IEEE Robotics & Automation Magazine.

[27]  Sebastian Scherer,et al.  A Principled Approach to Enable Safe and High Performance Maneuvers for Autonomous Rotorcraft , 2014 .

[28]  Charles Richter,et al.  High-speed autonomous navigation of unknown environments using learned probabilities of collision , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[29]  Nicholas Roy,et al.  Nonparametric Bayesian inference on multivariate exponential families , 2014, NIPS.

[30]  N. Roy,et al.  Markov Chain Hallway and Poisson Forest Environment Generating Distributions , 2015 .