Grasping POMDPs: Theory and Experiments

Abstract— We describe a method for planning under uncertainty for robotic manipulation of objects by partitioning the configuration space into a set of regions that are closed under compliant motions. These regions can be treated as states in a partially observable Markov decision process (POMDP), which can be solved to yield optimal control policies under uncertainty. We demonstrate the approach on simple grasping problems, showing that it can construct highly robust, efficiently executable solutions. In this paper we describe some initial experimentation with a Barrett arm and a Barret hand, instrumented with some contact sensors of our own design. This paper is based on [7].

[1]  Bruce Randall Donald,et al.  A Geometric Approach to Error Detection and Recovery for Robot Motion Planning with Uncertainty , 1987, Artif. Intell..

[2]  Sebastian Thrun,et al.  Coastal Navigation with Mobile Robots , 1999, NIPS.

[3]  Steven M. LaValle,et al.  Planning algorithms , 2006 .

[4]  Michael A. Erdmann,et al.  Using Backprojections for Fine Motion Planning with Uncertainty , 1986 .

[5]  S. LaValle,et al.  Motion Planning , 2008, Springer Handbook of Robotics.

[6]  Jean-Claude Latombe,et al.  Robot motion planning , 1970, The Kluwer international series in engineering and computer science.

[7]  Leslie Pack Kaelbling,et al.  Grasping POMDPs , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[8]  Joelle Pineau,et al.  Point-based value iteration: An anytime algorithm for POMDPs , 2003, IJCAI.

[9]  Roderic A. Grupen,et al.  Acquiring state from control dynamics to learn grasping policies for robot hands , 2002, Adv. Robotics.

[10]  Reid G. Simmons,et al.  Heuristic Search Value Iteration for POMDPs , 2004, UAI.

[11]  Russell H. Taylor,et al.  Automatic Synthesis of Fine-Motion Strategies for Robots , 1984 .

[12]  Edward J. Sondik,et al.  The Optimal Control of Partially Observable Markov Processes over a Finite Horizon , 1973, Oper. Res..

[13]  Robert Givan,et al.  Equivalence notions and model minimization in Markov decision processes , 2003, Artif. Intell..

[14]  Leslie Pack Kaelbling,et al.  Acting under uncertainty: discrete Bayesian models for mobile-robot navigation , 1996, Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems. IROS '96.

[15]  Alan D. Christiansen,et al.  Probabilistic Analysis of Manipulation Tasks: A Conceptual Framework , 1996, Int. J. Robotics Res..

[16]  Reid G. Simmons,et al.  Probabilistic Robot Navigation in Partially Observable Environments , 1995, IJCAI.

[17]  Jean-Claude Latombe,et al.  Robot Motion Planning with Uncertainty in Control and Sensing , 1991, Artif. Intell..

[18]  Jing Wei,et al.  Intelligent Strategies for Compliant Robotic Assembly , 2001 .

[19]  Doina Precup,et al.  Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..

[20]  Steven M. LaValle,et al.  An objective-based stochastic framework for manipulation planning , 1994, Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS'94).

[21]  H. Harry Asada,et al.  A discrete event approach to the control of robotic assembly tasks , 1993, [1993] Proceedings IEEE International Conference on Robotics and Automation.

[22]  Nikos A. Vlassis,et al.  Perseus: Randomized Point-based Value Iteration for POMDPs , 2005, J. Artif. Intell. Res..

[23]  Richard A. Volz,et al.  Acquiring robust, force-based assembly skills from human demonstration , 2000, IEEE Trans. Robotics Autom..