暂无分享,去创建一个
Pieter Abbeel | Igor Mordatch | Kevin Lu | Aditya Grover | P. Abbeel | Igor Mordatch | Kevin Lu | Aditya Grover
[1] A. M. Lyapunov. The general problem of the stability of motion , 1992 .
[2] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[3] Yishay Mansour,et al. Reinforcement Learning in POMDPs Without Resets , 2005, IJCAI.
[4] Chrystopher L. Nehaniv,et al. Empowerment: a universal agent-centric measure of control , 2005, 2005 IEEE Congress on Evolutionary Computation.
[5] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[6] Christoph Salge,et al. Empowerment - an Introduction , 2013, ArXiv.
[7] Evangelos Theodorou,et al. Model Predictive Path Integral Control using Covariance Variable Importance Sampling , 2015, ArXiv.
[8] Sergey Levine,et al. Learning compound multi-step controllers under unknown dynamics , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[9] Martial Hebert,et al. Improving Multi-Step Prediction of Learned Time Series Models , 2015, AAAI.
[10] Wojciech Zaremba,et al. OpenAI Gym , 2016, ArXiv.
[11] Doina Precup,et al. The Option-Critic Architecture , 2016, AAAI.
[12] MODEL-ENSEMBLE TRUST-REGION POLICY OPTI- , 2017 .
[13] Razvan Pascanu,et al. Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.
[14] Daan Wierstra,et al. Variational Intrinsic Control , 2016, ICLR.
[15] Pieter Abbeel,et al. Prediction and Control with Temporal Segment Models , 2017, ICML.
[16] Sergey Levine,et al. Uncertainty-Aware Reinforcement Learning for Collision Avoidance , 2017, ArXiv.
[17] Pieter Abbeel,et al. Model-Ensemble Trust-Region Policy Optimization , 2018, ICLR.
[18] Sergey Levine,et al. Leave no Trace: Learning to Reset for Safe and Autonomous Reinforcement Learning , 2017, ICLR.
[19] Sergey Levine,et al. Data-Efficient Hierarchical Reinforcement Learning , 2018, NeurIPS.
[20] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[21] Pieter Abbeel,et al. Variational Option Discovery Algorithms , 2018, ArXiv.
[22] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[23] Yee Whye Teh,et al. Progress & Compress: A scalable framework for continual learning , 2018, ICML.
[24] Sergey Levine,et al. Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[25] David Warde-Farley,et al. Unsupervised Control Through Non-Parametric Discriminative Rewards , 2018, ICLR.
[26] Pieter Abbeel,et al. Adaptive Online Planning for Continual Lifelong Learning , 2019, ArXiv.
[27] Yann LeCun,et al. Model-Predictive Policy Learning with Uncertainty Regularization for Driving in Dense Traffic , 2019, ICLR.
[28] Sergey Levine,et al. Deep Online Learning via Meta-Learning: Continual Adaptation for Model-Based RL , 2018, ICLR.
[29] Sergey Levine,et al. Meta-Learning , 2019, Automated Machine Learning.
[30] Dario Amodei,et al. Benchmarking Safe Exploration in Deep Reinforcement Learning , 2019 .
[31] Sergey Levine,et al. Why Does Hierarchy (Sometimes) Work So Well in Reinforcement Learning? , 2019, ArXiv.
[32] Sergey Levine,et al. Diversity is All You Need: Learning Skills without a Reward Function , 2018, ICLR.
[33] Erwan Lecarpentier,et al. Non-Stationary Markov Decision Processes a Worst-Case Approach using Model-Based Reinforcement Learning , 2019, NeurIPS.
[34] Sham M. Kakade,et al. Plan Online, Learn Offline: Efficient Learning and Exploration via Model-Based Control , 2018, ICLR.
[35] Patrick van der Smagt,et al. Unsupervised Real-Time Control Through Variational Empowerment , 2017, ISRR.
[36] Sergey Levine,et al. Deep Dynamics Models for Learning Dexterous Manipulation , 2019, CoRL.
[37] Katja Hofmann,et al. The MineRL Competition on Sample Efficient Reinforcement Learning using Human Priors , 2019, ArXiv.
[38] Sergey Levine,et al. When to Trust Your Model: Model-Based Policy Optimization , 2019, NeurIPS.
[39] Yifan Wu,et al. Behavior Regularized Offline Reinforcement Learning , 2019, ArXiv.
[40] David Rolnick,et al. Experience Replay for Continual Learning , 2018, NeurIPS.
[41] Ruben Villegas,et al. Learning Latent Dynamics for Planning from Pixels , 2018, ICML.
[42] Jordi Torres,et al. Explore, Discover and Learn: Unsupervised Discovery of State-Covering Skills , 2020, ICML.
[43] S. Levine,et al. Conservative Q-Learning for Offline Reinforcement Learning , 2020, NeurIPS.
[44] Sergey Levine,et al. The Ingredients of Real-World Robotic Reinforcement Learning , 2020, ICLR.
[45] T. Joachims,et al. MOReL : Model-Based Offline Reinforcement Learning , 2020, NeurIPS.
[46] Chelsea Finn,et al. Cautious Adaptation For Reinforcement Learning in Safety-Critical Settings , 2020, ICML.
[47] Rishabh Agarwal,et al. An Optimistic Perspective on Offline Reinforcement Learning , 2019, ICML.
[48] Pieter Abbeel,et al. Efficient Online Estimation of Empowerment for Reinforcement Learning , 2020, ArXiv.
[49] David Warde-Farley,et al. Fast Task Inference with Variational Intrinsic Successor Features , 2019, ICLR.
[50] S. Levine,et al. Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems , 2020, ArXiv.
[51] Doina Precup,et al. Options of Interest: Temporal Abstraction with Interest Functions , 2020, AAAI.
[52] Lantao Yu,et al. MOPO: Model-based Offline Policy Optimization , 2020, NeurIPS.
[53] Karol Hausman,et al. Emergent Real-World Robotic Skills via Unsupervised Off-Policy Reinforcement Learning , 2020, Robotics: Science and Systems.
[54] Sergey Levine,et al. Ecological Reinforcement Learning , 2020, ArXiv.
[55] Justin Fu,et al. D4RL: Datasets for Deep Data-Driven Reinforcement Learning , 2020, ArXiv.
[56] Jimmy Ba,et al. Exploring Model-based Planning with Policy Networks , 2019, ICLR.
[57] David Held,et al. Learning Off-Policy with Online Planning , 2020, CoRL.
[58] Sergey Levine,et al. Dynamics-Aware Unsupervised Discovery of Skills , 2019, ICLR.
[59] Gabriel Dulac-Arnold,et al. Model-Based Offline Planning , 2020, ICLR.
[60] Pieter Abbeel,et al. Efficient Empowerment Estimation for Unsupervised Stabilization , 2020, ICLR.