T-LQG: Closed-loop belief space planning via trajectory-optimized LQG

Planning under motion and observation uncertainties requires the solution of a stochastic control problem in the space of feedback policies. In this paper, by restricting the policy class to the linear feedback polices, we reduce the general (n2 + n)-dimensional belief space planning problem to an (n)-dimensional problem. As opposed to the previous literature that search in the space of open-loop optimal control policies, we obtain this reduction in the space of closed-loop policies by obtaining a Linear Quadratic Gaussian (LQG) design with the best nominal performance. Then, by taking the entire underlying trajectory of the LQG controller as the decision variable, we pose a coupled design of the trajectory and estimator (while keeping the design of the controller separate) as a NonLinear Program (NLP) that can be solved by a general NLP solver. We prove that under a first-order approximation and a careful usage of the separation principle, our approximations are valid. We provide an analysis on the existing major belief space planning methods and show that our algorithm keeps the lowest computational burden while searching in the policy space. Finally, we extend our solution to contain general state and control constraints. Our simulation results support our design.

[1]  Pieter Abbeel,et al.  Scaling up Gaussian Belief Space Planning Through Covariance-Free Trajectory Optimization and Automatic Differentiation , 2014, WAFR.

[2]  Panganamala Ramana Kumar,et al.  Feedback motion planning under non-Gaussian uncertainty and non-convex state constraints , 2015, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[3]  Joelle Pineau,et al.  Point-based value iteration: An anytime algorithm for POMDPs , 2003, IJCAI.

[4]  Ron Alterovitz,et al.  Motion planning under uncertainty using iterative local optimization in belief space , 2012, Int. J. Robotics Res..

[5]  Karl Johan Åström,et al.  Control: A perspective , 2014, Autom..

[6]  D. Bertsekas,et al.  Dynamic Programming and Stochastic Control , 1977, IEEE Transactions on Systems, Man, and Cybernetics.

[7]  Pravin Varaiya,et al.  Stochastic Systems: Estimation, Identification, and Adaptive Control , 1986 .

[8]  Panganamala Ramana Kumar,et al.  Belief Space Planning Simplified: Trajectory-Optimized LQG (T-LQG) (Extended Report) , 2016, ArXiv.

[9]  Edward J. Sondik,et al.  The Optimal Control of Partially Observable Markov Processes over a Finite Horizon , 1973, Oper. Res..

[10]  Ron Alterovitz,et al.  High-Frequency Replanning Under Uncertainty Using Parallel Sampling-Based Motion Planning , 2015, IEEE Transactions on Robotics.

[11]  Ron Alterovitz,et al.  Stochastic Extended LQR for Optimization-Based Motion Planning Under Uncertainty , 2016, IEEE Trans Autom. Sci. Eng..

[12]  Dimitri P. Bertsekas,et al.  Dynamic programming and optimal control, 3rd Edition , 2005 .

[13]  Leslie Pack Kaelbling,et al.  Belief space planning assuming maximum likelihood observations , 2010, Robotics: Science and Systems.

[14]  Surya P. N. Singh,et al.  An online and approximate solver for POMDPs with continuous action space , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[15]  Nancy M. Amato,et al.  FIRM: Sampling-based feedback motion-planning under motion uncertainty and imperfect measurements , 2014, Int. J. Robotics Res..

[16]  Siddhartha S. Srinivasa,et al.  CHOMP: Covariant Hamiltonian optimization for motion planning , 2013, Int. J. Robotics Res..

[17]  P. Abbeel,et al.  LQG-MP: Optimized path planning for robots with motion uncertainty and imperfect state information , 2011 .

[18]  Guy Shani,et al.  Noname manuscript No. (will be inserted by the editor) A Survey of Point-Based POMDP Solvers , 2022 .

[19]  Robert Platt,et al.  Convex Receding Horizon Control in Non-Gaussian Belief Space , 2012, WAFR.

[20]  N. Moshtagh MINIMUM VOLUME ENCLOSING ELLIPSOIDS , 2005 .