Integrating Human-Provided Information into Belief State Representation Using Dynamic Factorization

In partially observed environments, it can be useful for a human to provide the robot with declarative information that represents probabilistic relational constraints on properties of objects in the world, augmenting the robot's sensory observations. For instance, a robot tasked with a search-and-rescue mission may be informed by the human that two victims are probably in the same room. An important question arises: how should we represent the robot's internal knowledge so that this information is correctly processed and combined with raw sensory information? In this paper, we provide an efficient belief state representation that dynamically selects an appropriate factoring, combining aspects of the belief when they are correlated through information and separating them when they are not. This strategy works in open domains, in which the set of possible objects is not known in advance, and provides significant improvements in inference time over a static factoring, leading to more efficient planning for complex partially observed tasks. We validate our approach experimentally in two open-domain planning problems: a 2D discrete gridworld task and a 3D continuous cooking task. A supplementary video can be found at http://tinyurl.com/chitnis-iros-18.

[1]  Xavier Boyen,et al.  Tractable Inference for Complex Stochastic Processes , 1998, UAI.

[2]  C. Aitken,et al.  The logic of decision , 2014 .

[4]  Robert P. Goldman,et al.  Expressive Planning and Explicit Knowledge , 1996, AIPS.

[5]  K. Clint Slatton,et al.  Dynamic factor graphs: A novel framework for multiple features data fusion , 2010, 2010 IEEE International Conference on Acoustics, Speech and Signal Processing.

[6]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[7]  Matthias Scheutz,et al.  Planning for human-robot teaming in open worlds , 2010, TIST.

[8]  Pierre Lison,et al.  Belief modelling for situation awareness in human-robot interaction , 2010, 19th International Symposium in Robot and Human Interactive Communication.

[9]  Peter Stone,et al.  CORPP: Commonsense Reasoning and Probabilistic Planning, as Applied to Dialog with a Mobile Robot , 2015, AAAI.

[10]  Guy Shani,et al.  Replanning in Domains with Partial Information and Sensing Actions , 2011, IJCAI.

[11]  Subramanian Ramamoorthy,et al.  Exploiting Causality for Selective Belief Filtering in Dynamic Bayesian Networks (Extended Abstract) , 2014, IJCAI.

[12]  Pieter Abbeel,et al.  Combined task and motion planning through an extensible planner-independent interface layer , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[13]  Joelle Pineau,et al.  Point-based value iteration: An anytime algorithm for POMDPs , 2003, IJCAI.

[14]  Brian Sallans,et al.  Learning Factored Representations for Partially Observable Markov Decision Processes , 1999, NIPS.

[15]  Moshe Dor,et al.  אבן, and: Stone , 2017 .

[16]  Brendan J. Frey,et al.  Factor graphs and the sum-product algorithm , 2001, IEEE Trans. Inf. Theory.

[17]  Scott R. Sleep,et al.  An adaptive belief representation for target tracking using disparate sensors in Wireless Sensor Networks , 2013, Proceedings of the 16th International Conference on Information Fusion.

[18]  Blai Bonet,et al.  Belief Tracking for Planning with Sensing: Width, Complexity and Approximations , 2014, J. Artif. Intell. Res..

[19]  Wheeler Ruml,et al.  World Planning for Robots via Hindsight Optimization , 2013 .

[20]  Matthew Richardson,et al.  Markov logic networks , 2006, Machine Learning.

[21]  Subbarao Kambhampati,et al.  Probabilistic Planning via Determinization in Hindsight , 2008, AAAI.

[22]  Dylan Hadfield-Menell,et al.  Modular task and motion planning in belief space , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[23]  P ? ? ? ? ? ? ? % ? ? ? ? , 1991 .

[24]  David Hsu,et al.  SARSOP: Efficient Point-Based POMDP Planning by Approximating Optimally Reachable Belief Spaces , 2008, Robotics: Science and Systems.

[25]  Kevin P. Murphy,et al.  The Factored Frontier Algorithm for Approximate Inference in DBNs , 2001, UAI.

[26]  Nando de Freitas,et al.  An Introduction to MCMC for Machine Learning , 2004, Machine Learning.

[27]  John McCarthy,et al.  Programs with common sense , 1960 .

[28]  Joel Veness,et al.  Monte-Carlo Planning in Large POMDPs , 2010, NIPS.

[29]  David Hsu,et al.  DESPOT: Online POMDP Planning with Regularization , 2013, NIPS.

[30]  Peter S. Maybeck,et al.  Stochastic Models, Estimation And Control , 2012 .

[31]  Michael I. Jordan,et al.  Loopy Belief Propagation for Approximate Inference: An Empirical Study , 1999, UAI.

[32]  Matthias Scheutz,et al.  Architectural Mechanisms for Handling Human Instructions in Open-World Mixed-Initiative Team Tasks , 2013 .

[33]  Yann LeCun,et al.  Dynamic Factor Graphs for Time Series Modeling , 2009, ECML/PKDD.

[34]  Robert Givan,et al.  FF-Replan: A Baseline for Probabilistic Planning , 2007, ICAPS.

[35]  Blai Bonet,et al.  Planning with Incomplete Information as Heuristic Search in Belief Space , 2000, AIPS.

[36]  Sriraam Natarajan,et al.  Learning Probabilistic Logic Models with Human Advice , 2015, AAAI Spring Symposia.

[37]  Leslie Pack Kaelbling,et al.  Belief space planning assuming maximum likelihood observations , 2010, Robotics: Science and Systems.

[38]  Roman Barták,et al.  Constraint Processing , 2009, Encyclopedia of Artificial Intelligence.