暂无分享,去创建一个
Wolfram Burgard | Jingwei Zhang | Joschka Boedecker | Nicolai Dorka | Niklas Wetzel | W. Burgard | J. Boedecker | Jingwei Zhang | Nicolai Dorka | Niklas Wetzel | Wolfram Burgard
[1] Razvan Pascanu,et al. Learning to Navigate in Complex Environments , 2016, ICLR.
[2] Joshua B. Tenenbaum,et al. Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation , 2016, NIPS.
[3] Samuel Gershman,et al. Design Principles of the Hippocampal Cognitive Map , 2014, NIPS.
[4] Alex Graves,et al. Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.
[5] A. P. Hyper-parameters. Count-Based Exploration with Neural Density Models , 2017 .
[6] Marlos C. Machado,et al. Eigenoption Discovery through the Deep Successor Representation , 2017, ICLR.
[7] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[8] Marc Pollefeys,et al. Episodic Curiosity through Reachability , 2018, ICLR.
[9] Ion Stoica,et al. DDCO: Discovery of Deep Continuous Options for Robot Learning from Demonstrations , 2017, CoRL.
[10] Shane Legg,et al. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.
[11] Tom Schaul,et al. Unifying Count-Based Exploration and Intrinsic Motivation , 2016, NIPS.
[12] Alexei A. Efros,et al. Large-Scale Study of Curiosity-Driven Learning , 2018, ICLR.
[13] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[14] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[15] Wojciech Jaskowski,et al. ViZDoom: A Doom-based AI research platform for visual reinforcement learning , 2016, 2016 IEEE Conference on Computational Intelligence and Games (CIG).
[16] Doina Precup,et al. The Option-Critic Architecture , 2016, AAAI.
[17] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[18] Pierre-Yves Oudeyer,et al. How can we define intrinsic motivation , 2008 .
[19] Romain Laroche,et al. On Value Function Representation of Long Horizon Problems , 2018, AAAI.
[20] Marlos C. Machado,et al. Count-Based Exploration with the Successor Representation , 2018, AAAI.
[21] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[22] Tom Schaul,et al. Successor Features for Transfer in Reinforcement Learning , 2016, NIPS.
[23] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[24] Tom Schaul,et al. FeUdal Networks for Hierarchical Reinforcement Learning , 2017, ICML.
[25] Peter Dayan,et al. Improving Generalization for Temporal Difference Learning: The Successor Representation , 1993, Neural Computation.
[26] Martin A. Riedmiller,et al. Learning by Playing - Solving Sparse Reward Tasks from Scratch , 2018, ICML.
[27] Balaraman Ravindran,et al. Successor Options : An Option Discovery Algorithm for Reinforcement Learning , 2018 .
[28] Wolfram Burgard,et al. Deep reinforcement learning with successor features for navigation across similar environments , 2016, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[29] Jürgen Schmidhuber,et al. Formal Theory of Creativity, Fun, and Intrinsic Motivation (1990–2010) , 2010, IEEE Transactions on Autonomous Mental Development.
[30] David Budden,et al. Distributed Prioritized Experience Replay , 2018, ICLR.
[31] Tom Schaul,et al. Reinforcement Learning with Unsupervised Auxiliary Tasks , 2016, ICLR.
[32] Samuel Gershman,et al. Deep Successor Reinforcement Learning , 2016, ArXiv.
[33] Amos J. Storkey,et al. Exploration by Random Network Distillation , 2018, ICLR.
[34] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[35] Pierre-Yves Oudeyer,et al. How can we define intrinsic motivation , 2008 .
[36] Murray Shanahan,et al. Feature Control as Intrinsic Motivation for Hierarchical Reinforcement Learning , 2017, IEEE Transactions on Neural Networks and Learning Systems.
[37] Yuval Tassa,et al. DeepMind Control Suite , 2018, ArXiv.
[38] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[39] Shane Legg,et al. DeepMind Lab , 2016, ArXiv.