Exploring Parameter Space with Structured Noise for Meta-Reinforcement Learning
暂无分享,去创建一个
Yu Zheng | Hui Xu | Deqiang Ouyang | Chong Zhang | Jiaxing Wang | Jie Shao | Deqiang Ouyang | Jie Shao | Yu Zheng | Jiaxing Wang | Chong Zhang | Hui Xu | Chong Zhang | Jiaxing Wang | Yu Zheng | Jie Shao
[1] Marcin Andrychowicz,et al. Parameter Space Noise for Exploration , 2017, ICLR.
[2] Ronen I. Brafman,et al. R-MAX - A General Polynomial Time Algorithm for Near-Optimal Reinforcement Learning , 2001, J. Mach. Learn. Res..
[3] Hang Li,et al. Meta-SGD: Learning to Learn Quickly for Few Shot Learning , 2017, ArXiv.
[4] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[5] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[6] Thomas L. Griffiths,et al. Recasting Gradient-Based Meta-Learning as Hierarchical Bayes , 2018, ICLR.
[7] Thomas G. Dietterich. What is machine learning? , 2020, Archives of Disease in Childhood.
[8] Yi Sun,et al. Planning to Be Surprised: Optimal Bayesian Exploration in Dynamic Environments , 2011, AGI.
[9] Filip De Turck,et al. VIME: Variational Information Maximizing Exploration , 2016, NIPS.
[10] Shane Legg,et al. Noisy Networks for Exploration , 2017, ICLR.
[11] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[12] Bogdan Gabrys,et al. Metalearning: a survey of trends and technologies , 2013, Artificial Intelligence Review.
[13] Filip De Turck,et al. #Exploration: A Study of Count-Based Exploration for Deep Reinforcement Learning , 2016, NIPS.