暂无分享,去创建一个
Zeb Kurth-Nelson | Demis Hassabis | Matthew Botvinick | Alexander Lerchner | David P. Reichert | Malcolm Reynolds | Gavin Buttimore | Neil C. Rabinowitz | Neil Rabinowitz | Mary Cassin | Nicolas Porcel | Loic Matthey | Francis Song | Jane X. Wang | Michael King | Tina Zhu | Charlie Deck | Peter Choy | D. Hassabis | M. Botvinick | Z. Kurth-Nelson | Malcolm Reynolds | L. Matthey | Alexander Lerchner | Mary Cassin | Tina Zhu | Francis Song | Michael King | Nicolas Porcel | Charlie Deck | Peter Choy | Gavin Buttimore
[1] C A Nelson,et al. Learning to Learn , 2017, Encyclopedia of Machine Learning and Data Mining.
[2] Jonathan Baxter,et al. Theoretical Models of Learning to Learn , 1998, Learning to Learn.
[3] Marc G. Bellemare,et al. The Arcade Learning Environment: An Evaluation Platform for General Agents , 2012, J. Artif. Intell. Res..
[4] Thomas L. Griffiths,et al. Recasting Gradient-Based Meta-Learning as Hierarchical Bayes , 2018, ICLR.
[5] Sergey Levine,et al. Efficient Off-Policy Meta-Reinforcement Learning via Probabilistic Context Variables , 2019, ICML.
[6] David Silver,et al. Meta-Gradient Reinforcement Learning , 2018, NeurIPS.
[7] Tom Schaul,et al. Reinforcement Learning with Unsupervised Auxiliary Tasks , 2016, ICLR.
[8] Eduardo F. Morales,et al. An Introduction to Reinforcement Learning , 2011 .
[9] Zeb Kurth-Nelson,et al. Learning to reinforcement learn , 2016, CogSci.
[10] Pieter Abbeel,et al. Some Considerations on Learning to Explore via Meta-Reinforcement Learning , 2018, ICLR 2018.
[11] Jane X. Wang,et al. Meta-learning in natural and artificial intelligence , 2020, Current Opinion in Behavioral Sciences.
[12] Guy Lever,et al. Human-level performance in 3D multiplayer games with population-based reinforcement learning , 2018, Science.
[13] Ruslan Salakhutdinov,et al. Actor-Mimic: Deep Multitask and Transfer Reinforcement Learning , 2015, ICLR.
[14] Razvan Pascanu,et al. Stabilizing Transformers for Reinforcement Learning , 2019, ICML.
[15] Razvan Pascanu,et al. Distilling Policy Distillation , 2019, AISTATS.
[16] Sergey Levine,et al. Emergent Complexity and Zero-shot Transfer via Unsupervised Environment Design , 2020, NeurIPS.
[17] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[18] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[19] Demis Hassabis,et al. Mastering Atari, Go, chess and shogi by planning with a learned model , 2019, Nature.
[20] Sergey Levine,et al. Meta-World: A Benchmark and Evaluation for Multi-Task and Meta Reinforcement Learning , 2019, CoRL.
[21] Marwan Mattar,et al. Unity: A General Platform for Intelligent Agents , 2018, ArXiv.
[22] Max Jaderberg,et al. Population Based Training of Neural Networks , 2017, ArXiv.
[23] Joshua B. Tenenbaum,et al. Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.
[24] Michael O. Duff,et al. Design for an Optimal Probe , 2003, ICML.
[25] Wojciech M. Czarnecki,et al. Grandmaster level in StarCraft II using multi-agent reinforcement learning , 2019, Nature.
[26] Marek Wydmuch,et al. ViZDoom Competitions: Playing Doom From Pixels , 2018, IEEE Transactions on Games.
[27] Simon Carter,et al. Using Unity to Help Solve Intelligence , 2020, ArXiv.
[28] Taehoon Kim,et al. Quantifying Generalization in Reinforcement Learning , 2018, ICML.
[29] H. Francis Song,et al. V-MPO: On-Policy Maximum a Posteriori Policy Optimization for Discrete and Continuous Control , 2019, ICLR.
[30] W. J. Studden,et al. Theory Of Optimal Experiments , 1972 .
[31] Jane X. Wang,et al. Reinforcement Learning, Fast and Slow , 2019, Trends in Cognitive Sciences.