A Bayesian View on Motor Control and Planning

The problem of motion control and planning can be formulated as an optimization problem. In this paper we discuss an alternative view that casts the problem as one of probabilistic inference. In simple cases where the optimization problem can be solved analytically the inference view leads to equivalent solutions. However, when approximate methods are necessary to tackle the problem, the tight relation between optimization and probabilistic inference has fruitfully lead to a transfer of methods between both fields. Here we show that such a transfer is also possible in the realm of robotics. The general idea is that motion can be generated by fusing motion objectives (task constraints, goals, motion priors) by using probabilistic inference techniques. In realistic scenarios exact inference is infeasible (as is the analytic solution of the corresponding optimization problem) and the use of efficient approximate inference methods is a promising alternative to classical motion optimization methods. In this paper we first derive Bayesian control methods that are directly analogous to classical redundant motion rate control and optimal dynamic control (including operational space control). Then, by extending the probabilistic models to be Markovian models of the whole trajectory, we show that approximate probabilistic inference methods (message passing) efficiently compute solutions to trajectory optimization problems. Using Gaussian belief approximations and local linearization the algorithm becomes related to Differential Dynamic Programming (DDP) (aka. iterative Linear Quadratic Gaussian (iLQG)).

[1]  Yoshihiko Nakamura,et al.  Inverse kinematic solutions with singularity robustness for robot manipulator control , 1986 .

[2]  T. Yoshikawa,et al.  Task-Priority Based Redundancy Control of Robot Manipulators , 1987 .

[3]  Steven M. LaValle,et al.  RRT-connect: An efficient approach to single-query path planning , 2000, Proceedings 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation. Symposia Proceedings (Cat. No.00CH37065).

[4]  Tom Minka,et al.  Expectation Propagation for approximate Bayesian inference , 2001, UAI.

[5]  Tom Minka,et al.  A family of algorithms for approximate Bayesian inference , 2001 .

[6]  Stuart J. Russell,et al.  Dynamic bayesian networks: representation, inference and learning , 2002 .

[7]  Svetha Venkatesh,et al.  Policy Recognition in the Abstract Hidden Markov Model , 2002, J. Artif. Intell. Res..

[8]  William T. Freeman,et al.  Understanding belief propagation and its generalizations , 2003 .

[9]  William T. Freeman,et al.  Comparison of graph cuts with belief propagation for stereo, using identical MRF parameters , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[10]  Masayuki Inaba,et al.  Motion Planning for Humanoid Robots , 2003, ISRR.

[11]  Ronan Boulic,et al.  An inverse kinematics architecture enforcing an arbitrary number of strict priority levels , 2004, The Visual Computer.

[12]  Weiwei Li,et al.  Hierarchical optimal control of redundant biomechanical systems , 2004, The 26th Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[13]  Jun Nakanishi,et al.  A unifying methodology for the control of robotic systems , 2005, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[14]  Jun Nakanishi,et al.  A Unifying Framework for the Control of Robotic Systems , 2005 .

[15]  Marc Toussaint,et al.  Probabilistic inference for solving (PO) MDPs , 2006 .

[16]  Bart Selman,et al.  Sparse Message Passing Algorithms for Weighted Maximum Satisfiability , 2007 .

[17]  Marc Toussaint,et al.  Optimization of sequential attractor-based movement for compact behaviour generation , 2007, 2007 7th IEEE-RAS International Conference on Humanoid Robots.

[18]  Marc Toussaint,et al.  Probabilistic inference for structured planning in robotics , 2007, 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[19]  M. Toussaint Lecture Notes: Factor graphs and belief propagation , 2008 .

[20]  Marc Toussaint,et al.  Model-free reinforcement learning as mixture learning , 2009, ICML '09.

[21]  Marc Toussaint,et al.  Robot trajectory optimization using approximate inference , 2009, ICML '09.

[22]  Sethu Vijayakumar,et al.  Methods for Learning Control Policies from Variable-Constraint Demonstrations , 2010, From Motor Learning to Interaction Learning in Robots.

[23]  Olivier Sigaud,et al.  Learning Forward Models for the Operational Space Control of Redundant Robots , 2010, From Motor Learning to Interaction Learning in Robots.

[24]  Olivier Sigaud,et al.  From Motor Learning to Interaction Learning in Robots , 2010, From Motor Learning to Interaction Learning in Robots.