Robot plans execution for information gathering tasks with resources constraints

Partially observable Markov decision processes (POMDPs) have been widely used to model real world problems because of their abilities to capture uncertainty in states, actions and observations. In robotics, there are also constraints imposed on the problems, such as time constraints or resources constraints for executing actions. In this work, we seek to address the problems of planning in the presence of both uncertainty and constraints. Constrained POMDPs extend the general POMDPs by explicitly representing constraints in the goal conditions. The method we take in this paper is to use a translation-based approach to generate an MDP policy off-line, and apply value of information calculation on-line to stochastically select the observation action by taking into account of information they gain and their resource usage. This on-line selection scheme was evaluated in a number of scenarios and simulations, and the preliminary results show that our approach can achieve better performance compared to deterministic schemes.

[1]  Reid G. Simmons,et al.  Point-Based POMDP Algorithms: Improved Analysis and Implementation , 2005, UAI.

[2]  Michael L. Littman,et al.  Algorithms for Sequential Decision Making , 1996 .

[3]  Jesse Hoey,et al.  A Decision-Theoretic Approach to Task Assistance for Persons with Dementia , 2005, IJCAI.

[4]  Nick Hawes,et al.  Using Qualitative Spatial Relations for indirect object search , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[5]  Edward J. Sondik,et al.  The Optimal Control of Partially Observable Markov Processes over a Finite Horizon , 1973, Oper. Res..

[6]  Steve J. Young,et al.  Partially observable Markov decision processes for spoken dialog systems , 2007, Comput. Speech Lang..

[7]  Richard D. Braatz,et al.  Piecewise Linear Dynamic Programming for Constrained POMDPs , 2008, AAAI.

[8]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[9]  David Hsu,et al.  SARSOP: Efficient Point-Based POMDP Planning by Approximating Optimally Reachable Belief Spaces , 2008, Robotics: Science and Systems.

[10]  Reid G. Simmons,et al.  Heuristic Search Value Iteration for POMDPs , 2004, UAI.

[11]  John N. Tsitsiklis,et al.  The Complexity of Markov Decision Processes , 1987, Math. Oper. Res..

[12]  Kee-Eung Kim,et al.  Point-Based Value Iteration for Constrained POMDPs , 2011, IJCAI.

[13]  Minlue Wang,et al.  Monitoring plan execution in partially observable stochastic worlds , 2014 .

[14]  Leslie Pack Kaelbling,et al.  Acting Optimally in Partially Observable Stochastic Domains , 1994, AAAI.

[15]  Richard Dearden,et al.  Improving robot plans for information gathering tasks through execution monitoring , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[16]  R. Bellman,et al.  Dynamic Programming and Markov Processes , 1960 .

[17]  Jesse Hoey,et al.  SPUDD: Stochastic Planning using Decision Diagrams , 1999, UAI.