暂无分享,去创建一个
Samuel Gershman | Ardavan Saeedi | Tejas D. Kulkarni | Simanta Gautam | S. Gershman | Ardavan Saeedi | Simanta Gautam | A. Saeedi
[1] Terrence J. Sejnowski,et al. Temporal Difference Learning of Position Evaluation in the Game of Go , 1993, NIPS.
[2] Peter Dayan,et al. Improving Generalization for Temporal Difference Learning: The Successor Representation , 1993, Neural Computation.
[3] Jitendra Malik,et al. Normalized cuts and image segmentation , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[4] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[5] Andrew G. Barto,et al. Automatic Discovery of Subgoals in Reinforcement Learning using Diverse Density , 2001, ICML.
[6] Marc W. Howard,et al. A distributed representation of temporal context , 2002 .
[7] Shie Mannor,et al. Q-Cut - Dynamic Discovery of Sub-goals in Reinforcement Learning , 2002, ECML.
[8] Sridhar Mahadevan,et al. Recent Advances in Hierarchical Reinforcement Learning , 2003, Discret. Event Dyn. Syst..
[9] Manfred Huber,et al. Subgoal Discovery for Hierarchical Reinforcement Learning Using Learned Policies , 2003, FLAIRS.
[10] Shie Mannor,et al. Dynamic abstraction in reinforcement learning via clustering , 2004, ICML.
[11] Alicia P. Wolfe,et al. Identifying useful subgoals in reinforcement learning by local graph partitioning , 2005, ICML.
[12] Andrew G. Barto,et al. Skill Discovery in Continuous Reinforcement Learning Domains using Skill Chaining , 2009, NIPS.
[13] Jürgen Schmidhuber,et al. Formal Theory of Creativity, Fun, and Intrinsic Motivation (1990–2010) , 2010, IEEE Transactions on Autonomous Mental Development.
[14] Per B. Sederberg,et al. The Successor Representation and Temporal Context , 2012, Neural Computation.
[15] Ari Weinstein,et al. Model-based hierarchical reinforcement learning and human action control , 2014, Philosophical Transactions of the Royal Society B: Biological Sciences.
[16] Shalabh Bhatnagar,et al. Universal Option Models , 2014, NIPS.
[17] Samuel Gershman,et al. Design Principles of the Hippocampal Cognitive Map , 2014, NIPS.
[18] P. Dayan,et al. The algorithmic anatomy of model-based evaluation , 2014, Philosophical Transactions of the Royal Society B: Biological Sciences.
[19] Sergey Levine,et al. Incentivizing Exploration In Reinforcement Learning With Deep Predictive Models , 2015, ArXiv.
[20] Jürgen Schmidhuber,et al. Binding via Reconstruction Clustering , 2015, ArXiv.
[21] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.
[22] Joshua B. Tenenbaum,et al. Deep Convolutional Inverse Graphics Network , 2015, NIPS.
[23] Kevin Murphy,et al. Efficient inference in occlusion-aware generative models of images , 2015, ArXiv.
[24] Wulfram Gerstner,et al. Attractor Network Dynamics Enable Preplay and Rapid Path Planning in Maze-like Environments , 2015, NIPS.
[25] Alex Graves,et al. DRAW: A Recurrent Neural Network For Image Generation , 2015, ICML.
[26] Shakir Mohamed,et al. Variational Information Maximisation for Intrinsically Motivated Reinforcement Learning , 2015, NIPS.
[27] Rob Fergus,et al. MazeBase: A Sandbox for Learning from Games , 2015, ArXiv.
[28] Shane Legg,et al. Massively Parallel Methods for Deep Reinforcement Learning , 2015, ArXiv.
[29] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[30] Daan Wierstra,et al. One-Shot Generalization in Deep Generative Models , 2016, ICML.
[31] Joshua B. Tenenbaum,et al. Understanding Visual Concepts with Continuation Learning , 2016, ArXiv.
[32] Geoffrey E. Hinton,et al. Attend, Infer, Repeat: Fast Scene Understanding with Generative Models , 2016, NIPS.
[33] Paolo Favaro,et al. Unsupervised Learning of Visual Representations by Solving Jigsaw Puzzles , 2016, ECCV.
[34] Marlos C. Machado,et al. Learning Purposeful Behaviour in the Absence of Rewards , 2016, ArXiv.
[35] Benjamin Van Roy,et al. Deep Exploration via Bootstrapped DQN , 2016, NIPS.
[36] Alex Graves,et al. Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.
[37] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[38] Wojciech Jaskowski,et al. ViZDoom: A Doom-based AI research platform for visual reinforcement learning , 2016, 2016 IEEE Conference on Computational Intelligence and Games (CIG).
[39] Tom Schaul,et al. Prioritized Experience Replay , 2015, ICLR.
[40] Joshua B. Tenenbaum,et al. Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation , 2016, NIPS.