暂无分享,去创建一个
[1] R. French. Catastrophic Forgetting in Connectionist Networks , 2006 .
[2] David Filliat,et al. State Representation Learning for Control: An Overview , 2018, Neural Networks.
[3] Joel Z. Leibo,et al. Unsupervised Predictive Memory in a Goal-Directed Agent , 2018, ArXiv.
[4] David Filliat,et al. S-RL Toolbox: Environments, Datasets and Evaluation Metrics for State Representation Learning , 2018, ArXiv.
[5] David Filliat,et al. Flatland: a Lightweight First-Person 2-D Environment for Reinforcement Learning , 2018, ArXiv.
[6] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[7] Guillaume Desjardins,et al. Understanding disentangling in $\beta$-VAE , 2018, 1804.03599.
[8] David Filliat,et al. Generative Models from the perspective of Continual Learning , 2018, 2019 International Joint Conference on Neural Networks (IJCNN).
[9] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[10] Christopher Burgess,et al. DARLA: Improving Zero-Shot Transfer in Reinforcement Learning , 2017, ICML.
[11] Bogdan Raducanu,et al. Memory Replay GANs: learning to generate images from new categories without forgetting , 2018, NeurIPS.
[12] David Filliat,et al. Decoupling feature extraction from policy learning: assessing benefits of state representation learning in goal based robotics , 2018, ArXiv.
[13] Jan Peters,et al. Stable reinforcement learning with autoencoders for tactile and visual data , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[14] Richard E. Turner,et al. Variational Continual Learning , 2017, ICLR.
[15] Christopher Burgess,et al. beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.
[16] Doina Precup,et al. Using Options for Knowledge Transfer in Reinforcement Learning , 1999 .
[17] Pascal Vincent,et al. Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[18] Jiwon Kim,et al. Continual Learning with Deep Generative Replay , 2017, NIPS.
[19] Tom Eccles,et al. Life-Long Disentangled Representation Learning with Cross-Domain Latent Homologies , 2018, NeurIPS.
[20] Guillaume Desjardins,et al. Understanding disentangling in β-VAE , 2018, ArXiv.
[21] Alex Graves,et al. Associative Compression Networks for Representation Learning , 2018, ArXiv.
[22] Peter Stone,et al. Transfer Learning for Reinforcement Learning Domains: A Survey , 2009, J. Mach. Learn. Res..
[23] Jürgen Schmidhuber,et al. Recurrent World Models Facilitate Policy Evolution , 2018, NeurIPS.
[24] Welch Bl. THE GENERALIZATION OF ‘STUDENT'S’ PROBLEM WHEN SEVERAL DIFFERENT POPULATION VARLANCES ARE INVOLVED , 1947 .
[25] Samy Bengio,et al. Generating Sentences from a Continuous Space , 2015, CoNLL.
[26] Alex Graves,et al. Associative Compression Networks , 2018 .