TAPIR: A software toolkit for approximating and adapting POMDP solutions online

One of the fundamental challenges in the design of autonomous robots is to reliably compute motion strategies in spite of significant uncertainty about sensor reliability, control errors, and unpredicatable events. The Partially Observable Markov Decision Process (POMDP) is a general and mathematically principled framework for this type of problem. Although exact solutions are computationally intractable, modern approximate POMDP solvers have made POMDP-based approaches practical for robotics tasks. However, almost all existing POMDP-based planning software suffers from at least two major issues. Firstly, most POMDP solvers require the model to be known a priori and remain constant during runtime, and secondly, quite a lot of the existing software is not very userfriendly. This paper presents the Toolkit for approximating and Adapting POMDP solutions In Real time (TAPIR), which tackles both problems. The need for a constant, fully known POMDP model is averted by implementing the recent Adaptive Belief Tree (ABT) algorithm, while user-friendliness is ensured by a welldocumented modular design, which also includes interfaces for the commonly-used Robotics Operating System (ROS) framework, and the high fidelity simulator V-REP. TAPIR can be downloaded from http://robotics.itee.uq. edu.au/~tapir. To the best of our knowledge, TAPIR is the first software toolkit that directly addresses the aforementioned problems.

[1]  Joel Veness,et al.  Monte-Carlo Planning in Large POMDPs , 2010, NIPS.

[2]  Nicholas M. Patrikalakis,et al.  Global motion planning under uncertain motion, sensing, and environment map , 2011, Auton. Robots.

[3]  Sebastian Thrun,et al.  Monte Carlo POMDPs , 1999, NIPS.

[4]  David Hsu,et al.  SARSOP: Efficient Point-Based POMDP Planning by Approximating Optimally Reachable Belief Spaces , 2008, Robotics: Science and Systems.

[5]  Hanna Kurniawati,et al.  An Online POMDP Solver for Uncertainty Planning in Dynamic Environment , 2013, ISRR.

[6]  Nicholas M. Patrikalakis,et al.  Point-Based Policy Transformation: Adapting Policy to Changing POMDP Models , 2012, WAFR.

[7]  Kris K. Hauser,et al.  Randomized Belief-Space Replanning in Partially-Observable Continuous Spaces , 2010, WAFR.

[8]  Mykel J. Kochenderfer,et al.  Unmanned Aircraft Collision Avoidance using Continuous-State POMDPs , 2011, Robotics: Science and Systems.

[9]  Reid G. Simmons,et al.  Heuristic Search Value Iteration for POMDPs , 2004, UAI.

[10]  Siddhartha S. Srinivasa,et al.  Pre- and post-contact policy decomposition for planar contact manipulation under uncertainty , 2014, Int. J. Robotics Res..

[11]  Csaba Szepesvári,et al.  Bandit Based Monte-Carlo Planning , 2006, ECML.

[12]  Joel W. Burdick,et al.  Interactive non-prehensile manipulation for grasping via POMDPs , 2013, 2013 IEEE International Conference on Robotics and Automation.

[13]  Reid G. Simmons,et al.  Point-Based POMDP Algorithms: Improved Analysis and Implementation , 2005, UAI.

[14]  David Hsu,et al.  Monte Carlo Value Iteration for Continuous-State POMDPs , 2010, WAFR.

[15]  Nicholas Roy,et al.  Global Motion Planning under Uncertain Motion, Sensing, and Environment Map , 2012 .

[16]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[17]  Nicholas Roy,et al.  Unmanned Aircraft Collision Avoidance Using Continuous-State POMDPs , 2012 .

[18]  Nikos A. Vlassis,et al.  Perseus: Randomized Point-based Value Iteration for POMDPs , 2005, J. Artif. Intell. Res..

[19]  Joelle Pineau,et al.  Point-based value iteration: An anytime algorithm for POMDPs , 2003, IJCAI.

[20]  Leslie Pack Kaelbling,et al.  Collision Avoidance for Unmanned Aircraft using Markov Decision Processes , 2010 .

[21]  John N. Tsitsiklis,et al.  The Complexity of Markov Decision Processes , 1987, Math. Oper. Res..

[22]  Morgan Quigley,et al.  ROS: an open-source Robot Operating System , 2009, ICRA 2009.

[23]  Surya P. N. Singh,et al.  V-REP: A versatile and scalable robot simulation framework , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[24]  Nicholas Roy,et al.  PUMA: Planning Under Uncertainty with Macro-Actions , 2010, AAAI.

[25]  Wolfram Burgard,et al.  OctoMap: an efficient probabilistic 3D mapping framework based on octrees , 2013, Autonomous Robots.