Inferring Task Goals and Constraints using Bayesian Nonparametric Inverse Reinforcement Learning
暂无分享,去创建一个
Nicholas Roy | Subhro Roy | Michael Noseworthy | Rohan Paul | Daehyung Park | N. Roy | Rohan Paul | Michael Noseworthy | Subhro Roy | Daehyung Park
[1] Anind K. Dey,et al. Maximum Entropy Inverse Reinforcement Learning , 2008, AAAI.
[2] Aude Billard,et al. Task Parameterization Using Continuous Constraints Extracted From Human Demonstrations , 2015, IEEE Transactions on Robotics.
[3] S. LaValle. Rapidly-exploring random trees : a new tool for path planning , 1998 .
[4] Donald Geman,et al. Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images , 1984 .
[5] Sergey Levine,et al. Improvisation through Physical Understanding: Using Novel Objects as Tools with Visual Foresight , 2019, Robotics: Science and Systems.
[6] Weichao Zhou,et al. Safety-Aware Apprenticeship Learning , 2018, CAV.
[7] Sergey Levine,et al. Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[8] Marc Toussaint,et al. Differentiable Physics and Stable Modes for Tool-Use and Manipulation Planning , 2018, Robotics: Science and Systems.
[9] Andrew Y. Ng,et al. Pharmacokinetics of a novel formulation of ivermectin after administration to goats , 2000, ICML.
[10] Sergey Levine,et al. Guided Cost Learning: Deep Inverse Optimal Control via Policy Optimization , 2016, ICML.
[11] Donald J. Berndt,et al. Using Dynamic Time Warping to Find Patterns in Time Series , 1994, KDD Workshop.
[12] Pieter Abbeel,et al. Apprenticeship learning via inverse reinforcement learning , 2004, ICML.
[13] B. Faverjon,et al. Probabilistic Roadmaps for Path Planning in High-Dimensional Con(cid:12)guration Spaces , 1996 .
[14] Sergey Levine,et al. Learning Complex Dexterous Manipulation with Deep Reinforcement Learning and Demonstrations , 2017, Robotics: Science and Systems.
[15] Julie A. Shah,et al. C-LEARN: Learning geometric constraints from demonstrations for multi-step manipulation in shared autonomy , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[16] Aaron F. Bobick,et al. A State-Based Approach to the Representation and Recognition of Gesture , 1997, IEEE Trans. Pattern Anal. Mach. Intell..
[17] Dmitry Berenson,et al. Learning Object Orientation Constraints and Guiding Constraints for Narrow Passages from One Demonstration , 2016, ISER.
[18] Marc Toussaint,et al. Logic-Geometric Programming: An Optimization-Based Approach to Combined Task and Motion Planning , 2015, IJCAI.
[19] Scott Kuindersma,et al. Robot learning from demonstration by constructing skill trees , 2012, Int. J. Robotics Res..
[20] Jonathan P. How,et al. Bayesian Nonparametric Inverse Reinforcement Learning , 2012, ECML/PKDD.
[21] Leopoldo Armesto,et al. Efficient learning of constraints and generic null space policies , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[22] Mark H. Overmars,et al. The Gaussian sampling strategy for probabilistic roadmap planners , 1999, Proceedings 1999 IEEE International Conference on Robotics and Automation (Cat. No.99CH36288C).
[23] Marc Toussaint,et al. Direct Loss Minimization Inverse Optimal Control , 2015, Robotics: Science and Systems.
[24] Jan Peters,et al. Relative Entropy Inverse Reinforcement Learning , 2011, AISTATS.
[25] Ron Alterovitz,et al. Demonstration-Guided Motion Planning , 2011, ISRR.
[26] Pieter Abbeel,et al. Finding Locally Optimal, Collision-Free Trajectories with Sequential Convex Optimization , 2013, Robotics: Science and Systems.
[27] Byron Boots,et al. Graph-Based Inverse Optimal Control for Robot Manipulation , 2015, IJCAI.
[28] Oliver Kroemer,et al. Towards learning hierarchical skills for multi-phase manipulation tasks , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).
[29] J. Pitman. Combinatorial Stochastic Processes , 2006 .
[30] Jean-Claude Latombe,et al. On the Probabilistic Foundations of Probabilistic Roadmap Planning , 2006, Int. J. Robotics Res..
[31] Markus Wulfmeier,et al. Maximum Entropy Deep Inverse Reinforcement Learning , 2015, 1507.04888.
[32] Zackory M. Erickson,et al. Autonomous Tool Construction Using Part Shape and Attachment Prediction , 2019, Robotics: Science and Systems.
[33] Dmitry Berenson,et al. Learning constraints from demonstrations with grid and parametric representations , 2018, WAFR.
[34] Brett Browning,et al. A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..
[35] Stefano Ermon,et al. Generative Adversarial Imitation Learning , 2016, NIPS.