暂无分享,去创建一个
Amos J. Storkey | Antreas Antoniou | Harrison Edwards | A. Storkey | Harrison Edwards | Antreas Antoniou
[1] Amos J. Storkey,et al. Towards a Neural Statistician , 2016, ICLR.
[2] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[3] Sebastian Thrun,et al. Learning to Learn: Introduction and Overview , 1998, Learning to Learn.
[4] Geoffrey E. Hinton,et al. Using Fast Weights to Attend to the Recent Past , 2016, NIPS.
[5] Frank Hutter,et al. Fixing Weight Decay Regularization in Adam , 2017, ArXiv.
[6] Shie Mannor,et al. A Tutorial on the Cross-Entropy Method , 2005, Ann. Oper. Res..
[7] Joshua Achiam,et al. On First-Order Meta-Learning Algorithms , 2018, ArXiv.
[8] G. Evans,et al. Learning to Optimize , 2008 .
[9] Theodore Lim,et al. SMASH: One-Shot Model Architecture Search through HyperNetworks , 2017, ICLR.
[10] Quoc V. Le,et al. Neural Architecture Search with Reinforcement Learning , 2016, ICLR.
[11] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[12] Marcin Andrychowicz,et al. Learning to learn by gradient descent by gradient descent , 2016, NIPS.
[13] Joshua B. Tenenbaum,et al. Human-level concept learning through probabilistic program induction , 2015, Science.
[14] Frank Hutter,et al. Decoupled Weight Decay Regularization , 2017, ICLR.
[15] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[16] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[17] R. Rubinstein. The Cross-Entropy Method for Combinatorial and Continuous Optimization , 1999 .
[18] Amos J. Storkey,et al. Data Augmentation Generative Adversarial Networks , 2017, ICLR 2018.
[19] Rich Caruana,et al. Learning Many Related Tasks at the Same Time with Backpropagation , 1994, NIPS.
[20] Gregory Shakhnarovich,et al. FractalNet: Ultra-Deep Neural Networks without Residuals , 2016, ICLR.
[21] Hang Li,et al. Meta-SGD: Learning to Learn Quickly for Few Shot Learning , 2017, ArXiv.
[22] Frank Hutter,et al. SGDR: Stochastic Gradient Descent with Warm Restarts , 2016, ICLR.
[23] Zeb Kurth-Nelson,et al. Learning to reinforcement learn , 2016, CogSci.
[24] Mubarak Shah,et al. Task Agnostic Meta-Learning for Few-Shot Learning , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Aleksander Madry,et al. How Does Batch Normalization Help Optimization? (No, It Is Not About Internal Covariate Shift) , 2018, NIPS 2018.
[26] Ricardo Vilalta,et al. A Perspective View and Survey of Meta-Learning , 2002, Artificial Intelligence Review.
[27] Hugo Larochelle,et al. Optimization as a Model for Few-Shot Learning , 2016, ICLR.
[28] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).