Feedback motion planning under non-Gaussian uncertainty and non-convex state constraints
暂无分享,去创建一个
Panganamala Ramana Kumar | Amirhossein Tamjidi | Suman Chakravorty | Mohammadhussein Rafieisakhaei | P. Kumar | S. Chakravorty | A. Tamjidi | Mohammadhussein Rafieisakhaei
[1] Edward J. Sondik,et al. The optimal control of par-tially observable Markov processes , 1971 .
[2] David Hsu,et al. Monte Carlo Value Iteration with Macro-Actions , 2011, NIPS.
[3] Arnaud Doucet,et al. A survey of convergence results on particle filtering methods for practitioners , 2002, IEEE Trans. Signal Process..
[4] Karl Johan Åström,et al. Optimal control of Markov processes with incomplete state information , 1965 .
[5] Nancy M. Amato,et al. FIRM: Sampling-based feedback motion-planning under motion uncertainty and imperfect measurements , 2014, Int. J. Robotics Res..
[6] John N. Tsitsiklis,et al. The Complexity of Markov Decision Processes , 1987, Math. Oper. Res..
[7] Leslie Pack Kaelbling,et al. Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..
[8] R. Wets,et al. L-SHAPED LINEAR PROGRAMS WITH APPLICATIONS TO OPTIMAL CONTROL AND STOCHASTIC PROGRAMMING. , 1969 .
[9] Eric A. Hansen,et al. An Improved Grid-Based Approximation Algorithm for POMDPs , 2001, IJCAI.
[10] Masahiro Ono,et al. A Probabilistic Particle-Control Approximation of Chance-Constrained Stochastic Predictive Control , 2010, IEEE Transactions on Robotics.
[11] Timothy J. Robinson,et al. Sequential Monte Carlo Methods in Practice , 2003 .
[12] Nicholas Roy,et al. The Belief Roadmap: Efficient Planning in Linear POMDPs by Factoring the Covariance , 2007, ISRR.
[13] Anne Condon,et al. On the Undecidability of Probabilistic Planning and Infinite-Horizon Partially Observable Markov Decision Problems , 1999, AAAI/IAAI.
[14] Joel W. Burdick,et al. Robotic motion planning in dynamic, cluttered, uncertain environments , 2010, 2010 IEEE International Conference on Robotics and Automation.
[15] Leslie Pack Kaelbling,et al. Efficient Planning in Non-Gaussian Belief Spaces and Its Application to Robot Grasping , 2011, ISRR.
[16] Leslie Pack Kaelbling,et al. Belief space planning assuming maximum likelihood observations , 2010, Robotics: Science and Systems.
[17] M. Sniedovich. Dijkstra's algorithm revisited: the dynamic programming connexion , 2006 .
[18] C. Tomlin,et al. Closed-loop belief space planning for linear, Gaussian systems , 2011, 2011 IEEE International Conference on Robotics and Automation.
[19] Nancy M. Amato,et al. FIRM: Feedback controller-based information-state roadmap - A framework for motion planning under uncertainty , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[20] G. Roberts,et al. Monte Carlo Maximum Likelihood Estimation for Discretely Observed Diffusion Processes , 2009, 0903.0290.
[21] Joelle Pineau,et al. Point-based value iteration: An anytime algorithm for POMDPs , 2003, IJCAI.
[22] Ron Alterovitz,et al. Motion planning under uncertainty using iterative local optimization in belief space , 2012, Int. J. Robotics Res..
[23] Sebastian Thrun,et al. Probabilistic robotics , 2002, CACM.
[24] Pravin Varaiya,et al. Stochastic Systems: Estimation, Identification, and Adaptive Control , 1986 .
[25] Edward J. Sondik,et al. The Optimal Control of Partially Observable Markov Processes over a Finite Horizon , 1973, Oper. Res..
[26] John N. Tsitsiklis,et al. A survey of computational complexity results in systems and control , 2000, Autom..
[27] David Hsu,et al. Monte Carlo Value Iteration for Continuous-State POMDPs , 2010, WAFR.
[28] Dimitri P. Bertsekas,et al. Dynamic Programming and Optimal Control, Two Volume Set , 1995 .
[29] Guy Shani,et al. Noname manuscript No. (will be inserted by the editor) A Survey of Point-Based POMDP Solvers , 2022 .
[30] Robert Platt,et al. Convex Receding Horizon Control in Non-Gaussian Belief Space , 2012, WAFR.
[31] Nicholas Roy,et al. Rapidly-exploring Random Belief Trees for motion planning under uncertainty , 2011, 2011 IEEE International Conference on Robotics and Automation.