暂无分享,去创建一个
Abhinav Gupta | Kyunghyun Cho | William F. Whitney | Rajat Agarwal | William Whitney | A. Gupta | Kyunghyun Cho | Rajat Agarwal
[1] Sergey Levine,et al. QT-Opt: Scalable Deep Reinforcement Learning for Vision-Based Robotic Manipulation , 2018, CoRL.
[2] Martin A. Riedmiller,et al. Deep learning of visual control policies , 2010, ESANN.
[3] Sergey Levine,et al. EMI: Exploration with Mutual Information Maximizing State and Action Embeddings , 2018, ArXiv.
[4] Samuel Gershman,et al. Deep Successor Reinforcement Learning , 2016, ArXiv.
[5] Richard S. Sutton,et al. Dyna, an integrated architecture for learning, planning, and reacting , 1990, SGAR.
[6] Sergey Levine,et al. Self-Consistent Trajectory Autoencoder: Hierarchical Reinforcement Learning with Trajectory Embeddings , 2018, ICML.
[7] Herke van Hoof,et al. Addressing Function Approximation Error in Actor-Critic Methods , 2018, ICML.
[8] Naftali Tishby,et al. The information bottleneck method , 2000, ArXiv.
[9] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[10] Doina Precup,et al. The Option-Critic Architecture , 2016, AAAI.
[11] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[12] Tamim Asfour,et al. Model-Based Reinforcement Learning via Meta-Policy Optimization , 2018, CoRL.
[13] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[14] Ruben Villegas,et al. Learning Latent Dynamics for Planning from Pixels , 2018, ICML.
[15] Sergey Levine,et al. Latent Space Policies for Hierarchical Reinforcement Learning , 2018, ICML.
[16] Guy Lever,et al. Deterministic Policy Gradient Algorithms , 2014, ICML.
[17] M. Botvinick,et al. The hippocampus as a predictive map , 2016 .
[18] Tom Schaul,et al. Reinforcement Learning with Unsupervised Auxiliary Tasks , 2016, ICLR.
[19] Daan Wierstra,et al. Stochastic Backpropagation and Approximate Inference in Deep Generative Models , 2014, ICML.
[20] Marc G. Bellemare,et al. Safe and Efficient Off-Policy Reinforcement Learning , 2016, NIPS.
[21] M. Botvinick,et al. The successor representation in human reinforcement learning , 2016, Nature Human Behaviour.
[22] Pieter Abbeel,et al. Stochastic Neural Networks for Hierarchical Reinforcement Learning , 2016, ICLR.
[23] Karol Hausman,et al. Learning an Embedding Space for Transferable Robot Skills , 2018, ICLR.
[24] Yuval Tassa,et al. DeepMind Control Suite , 2018, ArXiv.
[25] Jan Peters,et al. Stable reinforcement learning with autoencoders for tactile and visual data , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[26] Joshua B. Tenenbaum,et al. Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation , 2016, NIPS.
[27] Martin A. Riedmiller,et al. PVEs: Position-Velocity Encoders for Unsupervised Learning of Structured State Representations , 2017, ArXiv.
[28] Soumith Chintala,et al. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.
[29] Shane Legg,et al. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.
[30] Christopher Burgess,et al. DARLA: Improving Zero-Shot Transfer in Reinforcement Learning , 2017, ICML.
[31] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[32] Sergey Levine,et al. Learning Actionable Representations with Goal-Conditioned Policies , 2018, ICLR.
[33] Sergey Levine,et al. Model-Based Reinforcement Learning for Atari , 2019, ICLR.
[34] Philip S. Thomas,et al. Learning Action Representations for Reinforcement Learning , 2019, ICML.
[35] Carl E. Rasmussen,et al. PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.
[36] Sergey Levine,et al. Data-Efficient Hierarchical Reinforcement Learning , 2018, NeurIPS.
[37] Richard Evans,et al. Deep Reinforcement Learning in Large Discrete Action Spaces , 2015, 1512.07679.
[38] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[39] Tom Schaul,et al. FeUdal Networks for Hierarchical Reinforcement Learning , 2017, ICML.
[40] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[41] Henry Zhu,et al. Soft Actor-Critic Algorithms and Applications , 2018, ArXiv.
[42] Sergey Levine,et al. EMI: Exploration with Mutual Information , 2018, ICML.
[43] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[44] Yann LeCun,et al. Model-Predictive Policy Learning with Uncertainty Regularization for Driving in Dense Traffic , 2019, ICLR.
[45] Jakub W. Pachocki,et al. Learning dexterous in-hand manipulation , 2018, Int. J. Robotics Res..
[46] Sergey Levine,et al. Near-Optimal Representation Learning for Hierarchical Reinforcement Learning , 2018, ICLR.
[47] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[48] Tom Schaul,et al. Successor Features for Transfer in Reinforcement Learning , 2016, NIPS.
[49] Alexander A. Alemi,et al. Deep Variational Information Bottleneck , 2017, ICLR.
[50] Peter Dayan,et al. Improving Generalization for Temporal Difference Learning: The Successor Representation , 1993, Neural Computation.
[51] Xiaodong Liu,et al. Cyclical Annealing Schedule: A Simple Approach to Mitigating KL Vanishing , 2019, NAACL.
[52] Shie Mannor,et al. The Natural Language of Actions , 2019, ICML.
[53] David Filliat,et al. Continual State Representation Learning for Reinforcement Learning using Generative Replay , 2018, ArXiv.
[54] Matthew W. Hoffman,et al. Distributed Distributional Deterministic Policy Gradients , 2018, ICLR.
[55] Christopher Burgess,et al. beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.