暂无分享,去创建一个
Pierre-Yves Oudeyer | Olivier Sigaud | Pierre Fournier | Cédric Colas | Olivier Sigaud | Cédric Colas | P. Oudeyer | Pierre Fournier
[1] Pierre-Yves Oudeyer,et al. Active learning of inverse models with intrinsically motivated goal exploration in robots , 2013, Robotics Auton. Syst..
[2] Pieter Abbeel,et al. Automatic Goal Generation for Reinforcement Learning Agents , 2017, ICML.
[3] Kate Saenko,et al. Hierarchical Reinforcement Learning with Hindsight , 2018, ArXiv.
[4] Leslie Pack Kaelbling,et al. Learning to Achieve Goals , 1993, IJCAI.
[5] Yee Whye Teh,et al. Distral: Robust multitask reinforcement learning , 2017, NIPS.
[6] Marcin Andrychowicz,et al. Overcoming Exploration in Reinforcement Learning with Demonstrations , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[7] Tom Schaul,et al. FeUdal Networks for Hierarchical Reinforcement Learning , 2017, ICML.
[8] Pierre-Yves Oudeyer,et al. Curiosity Driven Exploration of Learned Disentangled Goal Spaces , 2018, CoRL.
[9] Pierre-Yves Oudeyer,et al. The strategic student approach for life-long exploration and learning , 2012, 2012 IEEE International Conference on Development and Learning and Epigenetic Robotics (ICDL).
[10] Alex Graves,et al. Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.
[11] Pierre-Yves Oudeyer,et al. Intrinsically Motivated Goal Exploration Processes with Automatic Curriculum Learning , 2017, J. Mach. Learn. Res..
[12] H. B. Mann,et al. On a Test of Whether one of Two Random Variables is Stochastically Larger than the Other , 1947 .
[13] Tom Schaul,et al. Unicorn: Continual Learning with a Universal, Off-policy Agent , 2018, ArXiv.
[14] Pierre-Yves Oudeyer,et al. Active choice of teachers, learning strategies and goals for a socially guided intrinsic motivation learner , 2012, Paladyn J. Behav. Robotics.
[15] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[16] Pierre-Yves Oudeyer,et al. Maximizing Learning Progress: An Internal Reward System for Development , 2003, Embodied Artificial Intelligence.
[17] Martin A. Riedmiller,et al. Leveraging Demonstrations for Deep Reinforcement Learning on Robotics Problems with Sparse Rewards , 2017, ArXiv.
[18] Pierre-Yves Oudeyer,et al. The interaction of maturational constraints and intrinsic motivations in active motor development , 2011, 2011 IEEE International Conference on Development and Learning (ICDL).
[19] Herke van Hoof,et al. Addressing Function Approximation Error in Actor-Critic Methods , 2018, ICML.
[20] Wojciech Czarnecki,et al. Multi-task Deep Reinforcement Learning with PopArt , 2018, AAAI.
[21] Pierre-Yves Oudeyer,et al. Modular active curiosity-driven discovery of tool use , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[22] Martin A. Riedmiller,et al. Learning by Playing - Solving Sparse Reward Tasks from Scratch , 2018, ICML.
[23] Marcin Andrychowicz,et al. Multi-Goal Reinforcement Learning: Challenging Robotics Environments and Request for Research , 2018, ArXiv.
[24] Karl J. Friston,et al. A Bayesian Foundation for Individual Learning Under Uncertainty , 2011, Front. Hum. Neurosci..
[25] Shane Legg,et al. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.
[26] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[27] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.
[28] Jürgen Schmidhuber,et al. Curious model-building control systems , 1991, [Proceedings] 1991 IEEE International Joint Conference on Neural Networks.
[29] Pierre-Yves Oudeyer,et al. How Evolution May Work Through Curiosity-Driven Developmental Process , 2016, Top. Cogn. Sci..
[30] Sergey Levine,et al. Data-Efficient Hierarchical Reinforcement Learning , 2018, NeurIPS.
[31] Pierre-Yves Oudeyer,et al. GEP-PG: Decoupling Exploration and Exploitation in Deep Reinforcement Learning Algorithms , 2017, ICML.