暂无分享,去创建一个
Sergey Levine | Hyun Oh Song | Yeonwoo Jeong | Hyoungseok Kim | Jaekyeom Kim | S. Levine | Hyoungseok Kim | Jaekyeom Kim | Yeonwoo Jeong
[1] S. Varadhan,et al. Asymptotic evaluation of certain Markov process expectations for large time , 1975 .
[2] T. Kohonen,et al. Representation of Information in Spatial Maps which are Produced by Self-Organization , 1983 .
[3] Panu Somervuo,et al. Self-organizing maps of symbol strings , 1998, Neurocomputing.
[4] Andrew Y. Ng,et al. Policy Invariance Under Reward Transformations: Theory and Application to Reward Shaping , 1999, ICML.
[5] Yi Ma,et al. Robust principal component analysis? , 2009, JACM.
[6] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[7] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[8] Marc G. Bellemare,et al. Skip Context Tree Switching , 2014, ICML.
[9] Sergey Levine,et al. Incentivizing Exploration In Reinforcement Learning With Deep Predictive Models , 2015, ArXiv.
[10] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[11] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[12] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[13] Honglak Lee,et al. Action-Conditional Video Prediction using Deep Networks in Atari Games , 2015, NIPS.
[14] Marc G. Bellemare,et al. The Arcade Learning Environment: An Evaluation Platform for General Agents , 2012, J. Artif. Intell. Res..
[15] Pieter Abbeel,et al. Benchmarking Deep Reinforcement Learning for Continuous Control , 2016, ICML.
[16] Filip De Turck,et al. VIME: Variational Information Maximizing Exploration , 2016, NIPS.
[17] Pieter Abbeel,et al. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets , 2016, NIPS.
[18] Tom Schaul,et al. Unifying Count-Based Exploration and Intrinsic Motivation , 2016, NIPS.
[19] Sebastian Nowozin,et al. f-GAN: Training Generative Neural Samplers using Variational Divergence Minimization , 2016, NIPS.
[20] Filip De Turck,et al. #Exploration: A Study of Count-Based Exploration for Deep Reinforcement Learning , 2016, NIPS.
[21] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[22] Philippe Beaudoin,et al. Independently Controllable Factors , 2017, ArXiv.
[23] Justin Fu,et al. EX2: Exploration with Exemplar Models for Deep Reinforcement Learning , 2017, NIPS.
[24] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[25] Oriol Vinyals,et al. Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.
[26] Yoshua Bengio,et al. Mutual Information Neural Estimation , 2018, ICML.
[27] Yoshua Bengio,et al. Learning deep representations by mutual information estimation and maximization , 2018, ICLR.