Meta-Learning with Implicit Gradients
暂无分享,去创建一个
Sergey Levine | Chelsea Finn | Aravind Rajeswaran | Sham Kakade | S. Levine | S. Kakade | Chelsea Finn | A. Rajeswaran
[1] Walter Baur,et al. The Complexity of Partial Derivatives , 1983, Theor. Comput. Sci..
[2] Richard J. Mammone,et al. Meta-neural networks that learn by learning , 1992, [Proceedings 1992] IJCNN International Joint Conference on Neural Networks.
[3] A. Griewank. Some Bounds on the Complexity of Gradients, Jacobians, and Hessians , 1993 .
[4] Sebastian Thrun,et al. Learning to Learn , 1998, Springer US.
[5] Stephen J. Wright,et al. Numerical Optimization (Springer Series in Operations Research and Financial Engineering) , 2000 .
[6] Andreas Griewank,et al. Evaluating derivatives - principles and techniques of algorithmic differentiation, Second Edition , 2000, Frontiers in applied mathematics.
[7] Sepp Hochreiter,et al. Learning to Learn Using Gradient Descent , 2001, ICANN.
[8] Laurent Hascoët,et al. Enabling user-driven Checkpointing strategies in Reverse-mode Automatic Differentiation , 2006, ArXiv.
[9] Yurii Nesterov,et al. Cubic regularization of Newton method and its global performance , 2006, Math. Program..
[10] Chuan-Sheng Foo,et al. Efficient multiple hyperparameter learning for log-linear models , 2007, NIPS.
[11] James Martens,et al. Deep learning via Hessian-free optimization , 2010, ICML.
[12] Joshua B. Tenenbaum,et al. One shot learning of simple visual concepts , 2011, CogSci.
[13] Justin Domke,et al. Generic Methods for Optimization-Based Modeling , 2012, AISTATS.
[14] Sébastien Bubeck,et al. Convex Optimization: Algorithms and Complexity , 2014, Found. Trends Mach. Learn..
[15] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[16] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[17] Ryan P. Adams,et al. Gradient-based Hyperparameter Optimization through Reversible Learning , 2015, ICML.
[18] Gregory R. Koch,et al. Siamese Neural Networks for One-Shot Image Recognition , 2015 .
[19] Marcin Andrychowicz,et al. Learning to learn by gradient descent by gradient descent , 2016, NIPS.
[20] Fabian Pedregosa,et al. Hyperparameter optimization with approximate gradient , 2016, ICML.
[21] Peter L. Bartlett,et al. RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning , 2016, ArXiv.
[22] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[23] Bartunov Sergey,et al. Meta-Learning with Memory-Augmented Neural Networks , 2016 .
[24] Pieter Abbeel,et al. Meta-Learning with Temporal Convolutions , 2017, ArXiv.
[25] J. Zico Kolter,et al. OptNet: Differentiable Optimization as a Layer in Neural Networks , 2017, ICML.
[26] Zeb Kurth-Nelson,et al. Learning to reinforcement learn , 2016, CogSci.
[27] Hugo Larochelle,et al. Optimization as a Model for Few-Shot Learning , 2016, ICLR.
[28] Hong Yu,et al. Meta Networks , 2017, ICML.
[29] Sergey Levine,et al. One-Shot Visual Imitation Learning via Meta-Learning , 2017, CoRL.
[30] Richard S. Zemel,et al. Prototypical Networks for Few-shot Learning , 2017, NIPS.
[31] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[32] Sham M. Kakade,et al. Towards Generalization and Simplicity in Continuous Control , 2017, NIPS.
[33] Hang Li,et al. Meta-SGD: Learning to Learn Quickly for Few Shot Learning , 2017, ArXiv.
[34] Barak A. Pearlmutter,et al. Automatic differentiation in machine learning: a survey , 2015, J. Mach. Learn. Res..
[35] Michael I. Jordan,et al. How to Escape Saddle Points Efficiently , 2017, ICML.
[36] Jitendra Malik,et al. Learning to Optimize , 2016, ICLR.
[37] Paolo Frasconi,et al. Forward and Reverse Gradient-Based Hyperparameter Optimization , 2017, ICML.
[38] Leslie Pack Kaelbling,et al. Modular meta-learning , 2018, CoRL.
[39] Sergey Levine,et al. Meta-Learning and Universality: Deep Representations and Gradient Descent can Approximate any Learning Algorithm , 2017, ICLR.
[40] Bin Wu,et al. Deep Meta-Learning: Learning to Learn in the Concept Space , 2018, ArXiv.
[41] Sergey Levine,et al. Probabilistic Model-Agnostic Meta-Learning , 2018, NeurIPS.
[42] Igor Mordatch,et al. Concept Learning with Energy-Based Models , 2018, ICLR.
[43] Thomas L. Griffiths,et al. Recasting Gradient-Based Meta-Learning as Hierarchical Bayes , 2018, ICLR.
[44] Marco Pavone,et al. Meta-Learning Priors for Efficient Online Bayesian Regression , 2018, WAFR.
[45] Katja Hofmann,et al. CAML: Fast Context Adaptation via Meta-Learning , 2018, ArXiv.
[46] Pieter Abbeel,et al. A Simple Neural Attentive Meta-Learner , 2017, ICLR.
[47] Pieter Abbeel,et al. Continuous Adaptation via Meta-Learning in Nonstationary and Competitive Environments , 2017, ICLR.
[48] Joshua Achiam,et al. On First-Order Meta-Learning Algorithms , 2018, ArXiv.
[49] Sungwan Kim,et al. Auto-Meta: Automated Gradient Based Meta Learner Search , 2018, ArXiv.
[50] Alexandre Lacoste,et al. TADAM: Task dependent adaptive metric for improved few-shot learning , 2018, NeurIPS.
[51] Chelsea Finn,et al. Learning to Learn with Gradients , 2018 .
[52] Razvan Pascanu,et al. Meta-Learning with Latent Embedding Optimization , 2018, ICLR.
[53] Boi Faltings,et al. Meta-Learning for Low-resource Natural Language Generation in Task-oriented Dialogue Systems , 2019, IJCAI.
[54] Luca Bertinetto,et al. Meta-learning with differentiable closed-form solvers , 2018, ICLR.
[55] Katja Hofmann,et al. Fast Context Adaptation via Meta-Learning , 2018, ICML.
[56] Marco Pavone,et al. A Differentiable Augmented Lagrangian Method for Bilevel Nonlinear Optimization , 2019, Robotics: Science and Systems.
[57] Byron Boots,et al. Truncated Back-propagation for Bilevel Optimization , 2018, AISTATS.
[58] Joshua B. Tenenbaum,et al. Infinite Mixture Prototypes for Few-Shot Learning , 2019, ICML.
[59] Subhransu Maji,et al. Meta-Learning With Differentiable Convex Optimization , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[60] Hugo Larochelle,et al. Meta-Dataset: A Dataset of Datasets for Learning to Learn from Few Examples , 2019, ICLR.