Goal-Aware Prediction: Learning to Model What Matters
暂无分享,去创建一个
[1] Leslie Pack Kaelbling,et al. Learning to Achieve Goals , 1993, IJCAI.
[2] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[3] Lih-Yuan Deng,et al. The Cross-Entropy Method: A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation, and Machine Learning , 2006, Technometrics.
[4] Carl E. Rasmussen,et al. PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.
[5] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.
[6] Martin A. Riedmiller,et al. Embed to Control: A Locally Linear Latent Dynamics Model for Control from Raw Images , 2015, NIPS.
[7] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[8] C. Rasmussen,et al. Improving PILCO with Bayesian Neural Network Dynamics Models , 2016 .
[9] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[10] Jitendra Malik,et al. Learning to Poke by Poking: Experiential Learning of Intuitive Physics , 2016, NIPS.
[11] Sergey Levine,et al. End-to-End Training of Deep Visuomotor Policies , 2015, J. Mach. Learn. Res..
[12] Abhinav Gupta,et al. Supersizing self-supervision: Learning to grasp from 50K tries and 700 robot hours , 2015, 2016 IEEE International Conference on Robotics and Automation (ICRA).
[13] Sergey Levine,et al. Unsupervised Learning for Physical Interaction through Video Prediction , 2016, NIPS.
[14] Daniel Nikovski,et al. Value-Aware Loss Function for Model-based Reinforcement Learning , 2017, AISTATS.
[15] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[16] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[17] Vladlen Koltun,et al. Learning to Act by Predicting the Future , 2016, ICLR.
[18] Razvan Pascanu,et al. Imagination-Augmented Agents for Deep Reinforcement Learning , 2017, NIPS.
[19] Sergey Levine,et al. Deep visual foresight for planning robot motion , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[20] Sergey Levine,et al. Self-Supervised Visual Planning with Temporal Skip Connections , 2017, CoRL.
[21] Alberto Rodriguez,et al. Learning Synergies Between Pushing and Grasping with Self-Supervised Deep Reinforcement Learning , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[22] Sergey Levine,et al. Visual Reinforcement Learning with Imagined Goals , 2018, NeurIPS.
[23] Pieter Abbeel,et al. Learning Plannable Representations with Causal InfoGAN , 2018, NeurIPS.
[24] Misha Denil,et al. Learning Awareness Models , 2018, ICLR.
[25] Sergey Levine,et al. Visual Foresight: Model-Based Deep Reinforcement Learning for Vision-Based Robotic Control , 2018, ArXiv.
[26] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[27] Sergey Levine,et al. Stochastic Variational Video Prediction , 2017, ICLR.
[28] Byron Boots,et al. Differentiable MPC for End-to-end Planning and Control , 2018, NeurIPS.
[29] Sergey Levine,et al. SOLAR: Deep Structured Latent Representations for Model-Based Reinforcement Learning , 2018, ArXiv.
[30] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[31] Alexey Dosovitskiy,et al. End-to-End Driving Via Conditional Imitation Learning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[32] Sergey Levine,et al. Stochastic Adversarial Video Prediction , 2018, ArXiv.
[33] Rob Fergus,et al. Stochastic Video Generation with a Learned Prior , 2018, ICML.
[34] Sergey Levine,et al. QT-Opt: Scalable Deep Reinforcement Learning for Vision-Based Robotic Manipulation , 2018, CoRL.
[35] S. Levine,et al. Meta-World: A Benchmark and Evaluation for Multi-Task and Meta Reinforcement Learning , 2019, CoRL.
[36] Pieter Abbeel,et al. Learning Robotic Manipulation through Visual Planning and Acting , 2019, Robotics: Science and Systems.
[37] Li Fei-Fei,et al. Dynamics Learning with Cascaded Variational Inference for Multi-Step Manipulation , 2019, CoRL.
[38] Marc G. Bellemare,et al. DeepMDP: Learning Continuous Latent Space Models for Representation Learning , 2019, ICML.
[39] Aaron van den Oord,et al. Shaping Belief States with Generative Environment Models for RL , 2019, NeurIPS.
[40] Prabhat Nagarajan,et al. Learning Latent State Spaces for Planning through Reward Prediction , 2019, ArXiv.
[41] John D. Co-Reyes,et al. Entity Abstraction in Visual Model-Based Reinforcement Learning , 2019, CoRL.
[42] Martin A. Riedmiller,et al. Imagined Value Gradients: Model-Based Policy Optimization with Transferable Latent Dynamics Models , 2019, CoRL.
[43] Chelsea Finn,et al. Unsupervised Visuomotor Control through Distributional Planning Networks , 2019, Robotics: Science and Systems.
[44] Sergey Levine,et al. Deep Dynamics Models for Learning Dexterous Manipulation , 2019, CoRL.
[45] Sergey Levine,et al. When to Trust Your Model: Model-Based Policy Optimization , 2019, NeurIPS.
[46] Sergey Levine,et al. Goal-Conditioned Video Prediction , 2019 .
[47] Ruben Villegas,et al. High Fidelity Video Prediction with Large Stochastic Recurrent Neural Networks , 2019, NeurIPS.
[48] Sergey Levine,et al. Contextual Imagined Goals for Self-Supervised Robotic Learning , 2019, CoRL.
[49] Luke Metz,et al. Learning to Predict Without Looking Ahead: World Models Without Forward Prediction , 2019, NeurIPS.
[50] Sergey Levine,et al. Search on the Replay Buffer: Bridging Planning and Reinforcement Learning , 2019, NeurIPS.
[51] Sergey Levine,et al. Improvisation through Physical Understanding: Using Novel Objects as Tools with Visual Foresight , 2019, Robotics: Science and Systems.
[52] S. Levine,et al. RoboNet: Large-Scale Multi-Robot Learning , 2019, Conference on Robot Learning.
[53] Marco Pavone,et al. Robot Motion Planning in Learned Latent Spaces , 2018, IEEE Robotics and Automation Letters.
[54] Gregory D. Hager,et al. Visual Robot Task Planning , 2018, 2019 International Conference on Robotics and Automation (ICRA).
[55] Ruben Villegas,et al. Learning Latent Dynamics for Planning from Pixels , 2018, ICML.
[56] Jimmy Ba,et al. Dream to Control: Learning Behaviors by Latent Imagination , 2019, ICLR.
[57] Sergey Levine,et al. Model-Based Reinforcement Learning for Atari , 2019, ICLR.
[58] S. Levine,et al. Dynamical Distance Learning for Semi-Supervised and Unsupervised Skill Discovery , 2019, ICLR.
[59] Demis Hassabis,et al. Mastering Atari, Go, chess and shogi by planning with a learned model , 2019, Nature.
[60] Marcello Restelli,et al. Gradient-Aware Model-based Policy Search , 2019, AAAI.
[61] Chelsea Finn,et al. Hierarchical Foresight: Self-Supervised Learning of Long-Horizon Tasks via Visual Subgoal Generation , 2019, ICLR.
[62] S. Levine,et al. Safety Augmented Value Estimation From Demonstrations (SAVED): Safe Deep Model-Based RL for Sparse Cost Robotic Tasks , 2019, IEEE Robotics and Automation Letters.
[63] Sergey Levine,et al. Stochastic Latent Actor-Critic: Deep Reinforcement Learning with a Latent Variable Model , 2019, NeurIPS.
[64] Pieter Abbeel,et al. Hallucinative Topological Memory for Zero-Shot Visual Planning , 2020, ICML.
[65] Jakub W. Pachocki,et al. Learning dexterous in-hand manipulation , 2018, Int. J. Robotics Res..
[66] Jimmy Ba,et al. Exploring Model-based Planning with Policy Networks , 2019, ICLR.
[67] Roberto Calandra,et al. Objective Mismatch in Model-based Reinforcement Learning , 2020, L4DC.