暂无分享,去创建一个
Sergey Levine | Abhishek Gupta | Chelsea Finn | Benjamin Eysenbach | S. Levine | Abhishek Gupta | Chelsea Finn | Benjamin Eysenbach
[1] Richard J. Mammone,et al. Meta-neural networks that learn by learning , 1992, [Proceedings 1992] IJCNN International Joint Conference on Neural Networks.
[2] D. Wolpert,et al. No Free Lunch Theorems for Search , 1995 .
[3] Christopher G. Atkeson,et al. A comparison of direct and model-based reinforcement learning , 1997, Proceedings of International Conference on Robotics and Automation.
[4] Sepp Hochreiter,et al. Learning to Learn Using Gradient Descent , 2001, ICANN.
[5] L. D. Whitley,et al. Complexity Theory and the No Free Lunch Theorem , 2005 .
[6] Yoshua Bengio,et al. On the Optimization of a Synaptic Learning Rule , 2007 .
[7] Jürgen Schmidhuber,et al. Driven by Compression Progress: A Simple Principle Explains Essential Aspects of Subjective Beauty, Novelty, Surprise, Interestingness, Attention, Curiosity, Creativity, Art, Science, Music, Jokes , 2008, ABiALS.
[8] Carl E. Rasmussen,et al. PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.
[9] P. Cochat,et al. Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.
[10] Sergey Levine,et al. Incentivizing Exploration In Reinforcement Learning With Deep Predictive Models , 2015, ArXiv.
[11] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.
[12] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[13] Pieter Abbeel,et al. Benchmarking Deep Reinforcement Learning for Continuous Control , 2016, ICML.
[14] Filip De Turck,et al. VIME: Variational Information Maximizing Exploration , 2016, NIPS.
[15] Marcin Andrychowicz,et al. Learning to learn by gradient descent by gradient descent , 2016, NIPS.
[16] Benjamin Van Roy,et al. Deep Exploration via Bootstrapped DQN , 2016, NIPS.
[17] Tom Schaul,et al. Unifying Count-Based Exploration and Intrinsic Motivation , 2016, NIPS.
[18] Peter L. Bartlett,et al. RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning , 2016, ArXiv.
[19] Trevor Darrell,et al. Data-dependent Initializations of Convolutional Neural Networks , 2015, ICLR.
[20] Saiprasad Koturwar,et al. Weight Initialization of Deep Neural Networks(DNNs) using Data Statistics , 2017, ArXiv.
[21] Pierre-Yves Oudeyer,et al. Intrinsically Motivated Goal Exploration Processes with Automatic Curriculum Learning , 2017, J. Mach. Learn. Res..
[22] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[23] Zeb Kurth-Nelson,et al. Learning to reinforcement learn , 2016, CogSci.
[24] Hugo Larochelle,et al. Optimization as a Model for Few-Shot Learning , 2016, ICLR.
[25] Alex Graves,et al. Automated Curriculum Learning for Neural Networks , 2017, ICML.
[26] Shiliang Pu,et al. All You Need is Beyond a Good Init: Exploring Better Solution for Training Extremely Deep Convolutional Neural Networks with Orthonormality and Modulation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[28] Hong Yu,et al. Meta Networks , 2017, ICML.
[29] Sergey Levine,et al. One-Shot Visual Imitation Learning via Meta-Learning , 2017, CoRL.
[30] C A Nelson,et al. Learning to Learn , 2017, Encyclopedia of Machine Learning and Data Mining.
[31] Sergey Levine,et al. Deep visual foresight for planning robot motion , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[32] Li Zhang,et al. Learning to Learn: Meta-Critic Networks for Sample Efficient Learning , 2017, ArXiv.
[33] Daan Wierstra,et al. Variational Intrinsic Control , 2016, ICLR.
[34] Richard S. Zemel,et al. Prototypical Networks for Few-shot Learning , 2017, NIPS.
[35] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[36] Jitendra Malik,et al. Learning to Optimize , 2016, ICLR.
[37] Sergey Levine,et al. Self-Consistent Trajectory Autoencoder: Hierarchical Reinforcement Learning with Trajectory Embeddings , 2018, ICML.
[38] Sergey Levine,et al. Meta-Learning and Universality: Deep Representations and Gradient Descent can Approximate any Learning Algorithm , 2017, ICLR.
[39] Pieter Abbeel,et al. Automatic Goal Generation for Reinforcement Learning Agents , 2017, ICML.
[40] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[41] Pieter Abbeel,et al. Variational Option Discovery Algorithms , 2018, ArXiv.
[42] Sergey Levine,et al. Meta-Reinforcement Learning of Structured Exploration Strategies , 2018, NeurIPS.
[43] Jascha Sohl-Dickstein,et al. Learning Unsupervised Learning Rules , 2018, ArXiv.
[44] Pieter Abbeel,et al. A Simple Neural Attentive Meta-Learner , 2017, ICLR.
[45] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[46] Ilya Kostrikov,et al. Intrinsic Motivation and Automatic Curricula via Asymmetric Self-Play , 2017, ICLR.
[47] Sergey Levine,et al. Temporal Difference Models: Model-Free Deep RL for Model-Based Control , 2018, ICLR.
[48] Pieter Abbeel,et al. Continuous Adaptation via Meta-Learning in Nonstationary and Competitive Environments , 2017, ICLR.
[49] Allan Jabri,et al. Universal Planning Networks , 2018, ICML.
[50] Pieter Abbeel,et al. Some Considerations on Learning to Explore via Meta-Reinforcement Learning , 2018, ICLR 2018.
[51] Pieter Abbeel,et al. Evolved Policy Gradients , 2018, NeurIPS.
[52] Sergey Levine,et al. Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[53] David Warde-Farley,et al. Unsupervised Control Through Non-Parametric Discriminative Rewards , 2018, ICLR.
[54] Sergey Levine,et al. Unsupervised Learning via Meta-Learning , 2018, ICLR.
[55] Sergey Levine,et al. Efficient Off-Policy Meta-Reinforcement Learning via Probabilistic Context Variables , 2019, ICML.
[56] Allan Jabri,et al. Unsupervised Curricula for Visual Meta-Reinforcement Learning , 2019, NeurIPS.
[57] Sergey Levine,et al. Diversity is All You Need: Learning Skills without a Reward Function , 2018, ICLR.
[58] S. Levine,et al. Guided Meta-Policy Search , 2019, NeurIPS.
[59] Amos J. Storkey,et al. Assume, Augment and Learn: Unsupervised Few-Shot Meta-Learning via Random Labels and Data Augmentation , 2019, ArXiv.
[60] Sergey Levine,et al. Learning to Adapt in Dynamic, Real-World Environments through Meta-Reinforcement Learning , 2018, ICLR.
[61] Tamim Asfour,et al. ProMP: Proximal Meta-Policy Search , 2018, ICLR.
[62] Jascha Sohl-Dickstein,et al. Meta-Learning Update Rules for Unsupervised Representation Learning , 2018, ICLR.
[63] Sergey Levine,et al. Efficient Exploration via State Marginal Matching , 2019, ArXiv.
[64] Yijun Wang,et al. Learning to Transfer: Unsupervised Meta Domain Translation , 2019, ArXiv.
[65] Tiejun Huang,et al. Unsupervised Few-shot Learning via Self-supervised Training , 2019, ArXiv.
[66] Sergey Levine,et al. Dynamics-Aware Unsupervised Discovery of Skills , 2019, ICLR.