暂无分享,去创建一个
[1] Sergey Levine,et al. QT-Opt: Scalable Deep Reinforcement Learning for Vision-Based Robotic Manipulation , 2018, CoRL.
[2] Brijen Thananjeyan,et al. Safety Augmented Value Estimation From Demonstrations (SAVED): Safe Deep Model-Based RL for Sparse Cost Robotic Tasks , 2020, IEEE Robotics and Automation Letters.
[3] Demis Hassabis,et al. Mastering Atari, Go, chess and shogi by planning with a learned model , 2019, Nature.
[4] Jakub W. Pachocki,et al. Learning dexterous in-hand manipulation , 2018, Int. J. Robotics Res..
[5] Prabhat Nagarajan,et al. Learning Latent State Spaces for Planning through Reward Prediction , 2019, ArXiv.
[6] Chelsea Finn,et al. Hierarchical Foresight: Self-Supervised Learning of Long-Horizon Tasks via Visual Subgoal Generation , 2019, ICLR.
[7] Sergey Levine,et al. Visual Reinforcement Learning with Imagined Goals , 2018, NeurIPS.
[8] Jitendra Malik,et al. Learning to Poke by Poking: Experiential Learning of Intuitive Physics , 2016, NIPS.
[9] Jiajun Wu,et al. Entity Abstraction in Visual Model-Based Reinforcement Learning , 2019, CoRL.
[10] Ruben Villegas,et al. Learning Latent Dynamics for Planning from Pixels , 2018, ICML.
[11] Roberto Calandra,et al. Objective Mismatch in Model-based Reinforcement Learning , 2020, L4DC.
[12] Sergey Levine,et al. Stochastic Latent Actor-Critic: Deep Reinforcement Learning with a Latent Variable Model , 2019, NeurIPS.
[13] Luke Metz,et al. Learning to Predict Without Looking Ahead: World Models Without Forward Prediction , 2019, NeurIPS.
[14] Sergey Levine,et al. RoboNet: Large-Scale Multi-Robot Learning , 2019, CoRL.
[15] Sergey Levine,et al. Meta-World: A Benchmark and Evaluation for Multi-Task and Meta Reinforcement Learning , 2019, CoRL.
[16] Sergey Levine,et al. Self-Supervised Visual Planning with Temporal Skip Connections , 2017, CoRL.
[17] Sergey Levine,et al. End-to-End Training of Deep Visuomotor Policies , 2015, J. Mach. Learn. Res..
[18] Silvio Savarese,et al. Dynamics Learning with Cascaded Variational Inference for Multi-Step Manipulation , 2019, CoRL.
[19] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[20] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[21] Gregory D. Hager,et al. Visual Robot Task Planning , 2018, 2019 International Conference on Robotics and Automation (ICRA).
[22] Marc G. Bellemare,et al. DeepMDP: Learning Continuous Latent Space Models for Representation Learning , 2019, ICML.
[23] Aaron van den Oord,et al. Shaping Belief States with Generative Environment Models for RL , 2019, NeurIPS.
[24] C. Rasmussen,et al. Improving PILCO with Bayesian Neural Network Dynamics Models , 2016 .
[25] Alberto Rodriguez,et al. Learning Synergies Between Pushing and Grasping with Self-Supervised Deep Reinforcement Learning , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[26] Rob Fergus,et al. Stochastic Video Generation with a Learned Prior , 2018, ICML.
[27] Sergey Levine,et al. Model-Based Reinforcement Learning for Atari , 2019, ICLR.
[28] Sergey Levine,et al. When to Trust Your Model: Model-Based Policy Optimization , 2019, NeurIPS.
[29] Marco Pavone,et al. Robot Motion Planning in Learned Latent Spaces , 2018, IEEE Robotics and Automation Letters.
[30] Lih-Yuan Deng,et al. The Cross-Entropy Method: A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation, and Machine Learning , 2006, Technometrics.
[31] Alexey Dosovitskiy,et al. End-to-End Driving Via Conditional Imitation Learning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[32] Vladlen Koltun,et al. Learning to Act by Predicting the Future , 2016, ICLR.
[33] Leslie Pack Kaelbling,et al. Learning to Achieve Goals , 1993, IJCAI.
[34] Carl E. Rasmussen,et al. PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.
[35] Martin A. Riedmiller,et al. Imagined Value Gradients: Model-Based Policy Optimization with Transferable Latent Dynamics Models , 2019, CoRL.
[36] Sergey Levine,et al. Deep Dynamics Models for Learning Dexterous Manipulation , 2019, CoRL.
[37] Ruben Villegas,et al. High Fidelity Video Prediction with Large Stochastic Recurrent Neural Networks , 2019, NeurIPS.
[38] Jimmy Ba,et al. Exploring Model-based Planning with Policy Networks , 2019, ICLR.
[39] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.
[40] Razvan Pascanu,et al. Imagination-Augmented Agents for Deep Reinforcement Learning , 2017, NIPS.
[41] Chelsea Finn,et al. Unsupervised Visuomotor Control through Distributional Planning Networks , 2019, Robotics: Science and Systems.
[42] Sergey Levine,et al. Stochastic Adversarial Video Prediction , 2018, ArXiv.
[43] Pieter Abbeel,et al. Learning Robotic Manipulation through Visual Planning and Acting , 2019, Robotics: Science and Systems.
[44] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[45] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[46] Mohammad Norouzi,et al. Dream to Control: Learning Behaviors by Latent Imagination , 2019, ICLR.
[47] Sergey Levine,et al. Stochastic Variational Video Prediction , 2017, ICLR.
[48] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[49] Sergey Levine,et al. Improvisation through Physical Understanding: Using Novel Objects as Tools with Visual Foresight , 2019, Robotics: Science and Systems.
[50] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[51] Pieter Abbeel,et al. Hallucinative Topological Memory for Zero-Shot Visual Planning , 2020, ICML.
[52] Abhinav Gupta,et al. Supersizing self-supervision: Learning to grasp from 50K tries and 700 robot hours , 2015, 2016 IEEE International Conference on Robotics and Automation (ICRA).
[53] Marcello Restelli,et al. Gradient-Aware Model-based Policy Search , 2019, AAAI.
[54] Sergey Levine,et al. Unsupervised Learning for Physical Interaction through Video Prediction , 2016, NIPS.
[55] Sergey Levine,et al. Search on the Replay Buffer: Bridging Planning and Reinforcement Learning , 2019, NeurIPS.
[56] Martin A. Riedmiller,et al. Embed to Control: A Locally Linear Latent Dynamics Model for Control from Raw Images , 2015, NIPS.
[57] Byron Boots,et al. Differentiable MPC for End-to-end Planning and Control , 2018, NeurIPS.
[58] Sergey Levine,et al. Deep visual foresight for planning robot motion , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[59] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[60] Sergey Levine,et al. SOLAR: Deep Structured Latent Representations for Model-Based Reinforcement Learning , 2018, ArXiv.
[61] Sergey Levine,et al. Dynamical Distance Learning for Semi-Supervised and Unsupervised Skill Discovery , 2020, ICLR.
[62] Sergey Levine,et al. Contextual Imagined Goals for Self-Supervised Robotic Learning , 2019, CoRL.
[63] Daniel Nikovski,et al. Value-Aware Loss Function for Model-based Reinforcement Learning , 2017, AISTATS.
[64] Pieter Abbeel,et al. Learning Plannable Representations with Causal InfoGAN , 2018, NeurIPS.
[65] Misha Denil,et al. Learning Awareness Models , 2018, ICLR.
[66] Sergey Levine,et al. Goal-Conditioned Video Prediction , 2019 .
[67] Sergey Levine,et al. Visual Foresight: Model-Based Deep Reinforcement Learning for Vision-Based Robotic Control , 2018, ArXiv.