暂无分享,去创建一个
Arjun K. Bansal | Mariano Phielipp | Madhavun Candadai | Zach Dwiel | Mariano Phielipp | Zach Dwiel | Madhavun Candadai
[1] Andriy Mnih,et al. Disentangling by Factorising , 2018, ICML.
[2] Tom Schaul,et al. FeUdal Networks for Hierarchical Reinforcement Learning , 2017, ICML.
[3] Doina Precup,et al. The Option-Critic Architecture , 2016, AAAI.
[4] Pieter Abbeel,et al. Stochastic Neural Networks for Hierarchical Reinforcement Learning , 2016, ICLR.
[5] Geoffrey E. Hinton,et al. Feudal Reinforcement Learning , 1992, NIPS.
[6] Nando de Freitas,et al. Playing hard exploration games by watching YouTube , 2018, NeurIPS.
[7] Alex Graves,et al. Strategic Attentive Writer for Learning Macro-Actions , 2016, NIPS.
[8] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[9] J. Urgen Schmidhuber,et al. Neural sequence chunkers , 1991, Forschungsberichte, TU Munich.
[10] Christopher Burgess,et al. DARLA: Improving Zero-Shot Transfer in Reinforcement Learning , 2017, ICML.
[11] Roger B. Grosse,et al. Isolating Sources of Disentanglement in Variational Autoencoders , 2018, NeurIPS.
[12] Tom Schaul,et al. Reinforcement Learning with Unsupervised Auxiliary Tasks , 2016, ICLR.
[13] Sergey Levine,et al. Visual Reinforcement Learning with Imagined Goals , 2018, NeurIPS.
[14] Pieter Abbeel,et al. Meta Learning Shared Hierarchies , 2017, ICLR.
[15] Demis Hassabis,et al. Mastering the game of Go without human knowledge , 2017, Nature.
[16] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[17] Sergey Levine,et al. Data-Efficient Hierarchical Reinforcement Learning , 2018, NeurIPS.
[18] Christopher Burgess,et al. beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.
[19] Stuart J. Russell,et al. Reinforcement Learning with Hierarchies of Machines , 1997, NIPS.
[20] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[21] Kate Saenko,et al. Hierarchical Reinforcement Learning with Hindsight , 2018, ArXiv.
[22] Marcin Andrychowicz,et al. Overcoming Exploration in Reinforcement Learning with Demonstrations , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[23] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[24] Sergey Levine,et al. Near-Optimal Representation Learning for Hierarchical Reinforcement Learning , 2018, ICLR.
[25] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.