暂无分享,去创建一个
Sergey Levine | Ashvin Nair | Laura Smith | Vitchyr H. Pong | Catherine Huang | S. Levine | Laura Smith | Ashvin Nair | Catherine Huang
[1] Leslie Pack Kaelbling,et al. Learning to Achieve Goals , 1993, IJCAI.
[2] Michael I. Jordan,et al. Reinforcement Learning Algorithm for Partially Observable Markov Decision Problems , 1994, NIPS.
[3] Stefan Schaal,et al. Is imitation learning the route to humanoid robots? , 1999, Trends in Cognitive Sciences.
[4] Pieter Abbeel,et al. Apprenticeship learning via inverse reinforcement learning , 2004, ICML.
[5] Jonathan Kofman,et al. Teleoperation of a robot manipulator using a vision-based human-robot interface , 2005, IEEE Transactions on Industrial Electronics.
[6] J. Andrew Bagnell,et al. Efficient Reductions for Imitation Learning , 2010, AISTATS.
[7] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[8] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.
[9] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[10] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[11] Sergey Levine,et al. Guided Cost Learning: Deep Inverse Optimal Control via Policy Optimization , 2016, ICML.
[12] Samuel Gershman,et al. Deep Successor Reinforcement Learning , 2016, ArXiv.
[13] Stefano Ermon,et al. Generative Adversarial Imitation Learning , 2016, NIPS.
[14] Peter L. Bartlett,et al. RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning , 2016, ArXiv.
[15] Kyunghyun Cho,et al. End-to-End Goal-Driven Web Navigation , 2016, NIPS.
[16] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[17] Shane Legg,et al. Deep Reinforcement Learning from Human Preferences , 2017, NIPS.
[18] Yuval Tassa,et al. Emergence of Locomotion Behaviours in Rich Environments , 2017, ArXiv.
[19] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[20] Tom Schaul,et al. Successor Features for Transfer in Reinforcement Learning , 2016, NIPS.
[21] Henry Zhu,et al. Soft Actor-Critic Algorithms and Applications , 2018, ArXiv.
[22] Pierre-Yves Oudeyer,et al. GEP-PG: Decoupling Exploration and Exploitation in Deep Reinforcement Learning Algorithms , 2017, ICML.
[23] David Silver,et al. Meta-Gradient Reinforcement Learning , 2018, NeurIPS.
[24] Pierre-Yves Oudeyer,et al. Unsupervised Learning of Goal Spaces for Intrinsically Motivated Goal Exploration , 2018, ICLR.
[25] Sergey Levine,et al. Visual Reinforcement Learning with Imagined Goals , 2018, NeurIPS.
[26] Sergey Levine,et al. Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection , 2016, Int. J. Robotics Res..
[27] Julian Togelius,et al. Deep Reinforcement Learning for General Video Game AI , 2018, 2018 IEEE Conference on Computational Intelligence and Games (CIG).
[28] Karol Hausman,et al. Learning an Embedding Space for Transferable Robot Skills , 2018, ICLR.
[29] Sergey Levine,et al. Learning Robust Rewards with Adversarial Inverse Reinforcement Learning , 2017, ICLR 2017.
[30] Sergey Levine,et al. Meta-Reinforcement Learning of Structured Exploration Strategies , 2018, NeurIPS.
[31] Shane Legg,et al. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.
[32] Sergey Levine,et al. Temporal Difference Models: Model-Free Deep RL for Model-Based Control , 2018, ICLR.
[33] Sergey Levine,et al. Unsupervised Meta-Learning for Reinforcement Learning , 2018, ArXiv.
[34] Shane Legg,et al. Scalable agent alignment via reward modeling: a research direction , 2018, ArXiv.
[35] Tom Schaul,et al. Transfer in Deep Reinforcement Learning Using Successor Features and Generalised Policy Improvement , 2018, ICML.
[36] S. Levine,et al. Meta-World: A Benchmark and Evaluation for Multi-Task and Meta Reinforcement Learning , 2019, CoRL.
[37] Joelle Pineau,et al. Benchmarking Batch Deep Reinforcement Learning Algorithms , 2019, ArXiv.
[38] David Warde-Farley,et al. Unsupervised Control Through Non-Parametric Discriminative Rewards , 2018, ICLR.
[39] Sergey Levine,et al. REPLAB: A Reproducible Low-Cost Arm Benchmark Platform for Robotic Learning , 2019, ArXiv.
[40] Sergey Levine,et al. Efficient Off-Policy Meta-Reinforcement Learning via Probabilistic Context Variables , 2019, ICML.
[41] Allan Jabri,et al. Unsupervised Curricula for Visual Meta-Reinforcement Learning , 2019, NeurIPS.
[42] Raia Hadsell,et al. Disentangled Cumulants Help Successor Representations Transfer to New Tasks , 2019, ArXiv.
[43] Doina Precup,et al. Off-Policy Deep Reinforcement Learning without Exploration , 2018, ICML.
[44] Sergey Levine,et al. Stabilizing Off-Policy Q-Learning via Bootstrapping Error Reduction , 2019, NeurIPS.
[45] Rémi Munos,et al. Recurrent Experience Replay in Distributed Reinforcement Learning , 2018, ICLR.
[46] Yee Whye Teh,et al. Meta reinforcement learning as task inference , 2019, ArXiv.
[47] Yifan Wu,et al. Behavior Regularized Offline Reinforcement Learning , 2019, ArXiv.
[48] Tamim Asfour,et al. ProMP: Proximal Meta-Policy Search , 2018, ICLR.
[49] S. Levine,et al. Accelerating Online Reinforcement Learning with Offline Datasets , 2020, ArXiv.
[50] S. Levine,et al. Learning Agile Robotic Locomotion Skills by Imitating Animals , 2020, Robotics: Science and Systems.
[51] Nando de Freitas,et al. Semi-supervised reward learning for offline reinforcement learning , 2020, ArXiv.
[52] Misha Denil,et al. Offline Learning from Demonstrations and Unlabeled Experience , 2020, ArXiv.
[53] Luisa M. Zintgraf,et al. VariBAD: A Very Good Method for Bayes-Adaptive Deep RL via Meta-Learning , 2019, ICLR.
[54] Improving Generalization in Meta Reinforcement Learning using Learned Objectives , 2019, ICLR.
[55] Ludovic Denoyer,et al. Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization , 2020, ArXiv.
[56] Oleg O. Sushkov,et al. Scaling data-driven robotics with reward sketching and batch reinforcement learning , 2019, Robotics: Science and Systems.
[57] S. Levine,et al. Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems , 2020, ArXiv.
[58] Junhyuk Oh,et al. Meta-Gradient Reinforcement Learning with an Objective Discovered Online , 2020, NeurIPS.
[59] Aviv Tamar,et al. Offline Meta Reinforcement Learning , 2020, ArXiv.
[60] Alexei A. Efros,et al. Test-Time Training with Self-Supervision for Generalization under Distribution Shifts , 2019, ICML.
[61] Justin Fu,et al. D4RL: Datasets for Deep Data-Driven Reinforcement Learning , 2020, ArXiv.
[62] Sergey Levine,et al. MELD: Meta-Reinforcement Learning from Images via Latent State Models , 2020, CoRL.
[63] Anca D. Dragan,et al. SQIL: Imitation Learning via Reinforcement Learning with Sparse Rewards , 2019, ICLR.
[64] Misha Denil,et al. Positive-Unlabeled Reward Learning , 2019, CoRL.
[65] WebGPT: Browser-assisted question-answering with human feedback , 2021, ArXiv.
[66] Sergey Levine,et al. Offline Meta-Reinforcement Learning with Advantage Weighting , 2020, ICML.
[67] Martin A. Riedmiller,et al. Towards Real Robot Learning in the Wild: A Case Study in Bipedal Locomotion , 2021, CoRL.
[68] Sergey Levine,et al. What Can I Do Here? Learning New Skills by Imagining Visual Affordances , 2021, 2021 IEEE International Conference on Robotics and Automation (ICRA).