暂无分享,去创建一个
[1] Peter Stone,et al. Transfer Learning for Reinforcement Learning Domains: A Survey , 2009, J. Mach. Learn. Res..
[2] Doina Precup,et al. Representation Discovery for MDPs Using Bisimulation Metrics , 2015, AAAI.
[3] Andrew W. Moore,et al. Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..
[4] Peter Dayan,et al. Improving Generalization for Temporal Difference Learning: The Successor Representation , 1993, Neural Computation.
[5] Robert Givan,et al. Equivalence notions and model minimization in Markov decision processes , 2003, Artif. Intell..
[6] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[7] Richard S. Sutton,et al. Integrated Architectures for Learning, Planning, and Reacting Based on Approximating Dynamic Programming , 1990, ML.
[8] Tom Schaul,et al. Successor Features for Transfer in Reinforcement Learning , 2016, NIPS.
[9] Stefanie Tellex,et al. Advantages and Limitations of using Successor Features for Transfer in Reinforcement Learning , 2017, ArXiv.
[10] Radford M. Neal. Pattern Recognition and Machine Learning , 2007, Technometrics.
[11] Doina Precup,et al. Metrics for Finite Markov Decision Processes , 2004, AAAI.
[12] M. Botvinick,et al. The successor representation in human reinforcement learning , 2016, Nature Human Behaviour.
[13] Thomas J. Walsh,et al. Towards a Unified Theory of State Abstraction for MDPs , 2006, AI&M.
[14] Wolfram Burgard,et al. Deep reinforcement learning with successor features for navigation across similar environments , 2016, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[15] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.