暂无分享,去创建一个
[1] Sepp Hochreiter,et al. Learning to Learn Using Gradient Descent , 2001, ICANN.
[2] Yee Whye Teh,et al. Conditional Neural Processes , 2018, ICML.
[3] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[4] Hugo Larochelle,et al. Optimization as a Model for Few-Shot Learning , 2016, ICLR.
[5] Ambedkar Dukkipati,et al. Generative Adversarial Residual Pairwise Networks for One Shot Learning , 2017, ArXiv.
[6] Gavriel Salomon,et al. T RANSFER OF LEARNING , 1992 .
[7] Joshua B. Tenenbaum,et al. Human-level concept learning through probabilistic program induction , 2015, Science.
[8] Hung-Yu Tseng,et al. Regularizing Meta-Learning via Gradient Dropout , 2020, ACCV.
[9] Sergey Levine,et al. Meta-World: A Benchmark and Evaluation for Multi-Task and Meta Reinforcement Learning , 2019, CoRL.
[10] Taghi M. Khoshgoftaar,et al. A survey on Image Data Augmentation for Deep Learning , 2019, Journal of Big Data.
[11] Chris Pal,et al. Towards Understanding Generalization in Gradient-Based Meta-Learning , 2019, ArXiv.
[12] Alexander A. Alemi,et al. Deep Variational Information Bottleneck , 2017, ICLR.
[13] Fei Chao,et al. Task Augmentation by Rotating for Meta-Learning , 2020, ArXiv.
[14] Pieter Abbeel,et al. Reinforcement Learning with Augmented Data , 2020, NeurIPS.
[15] Sergey Levine,et al. One-Shot Visual Imitation Learning via Meta-Learning , 2017, CoRL.
[16] Hugo Larochelle,et al. Meta-Dataset: A Dataset of Datasets for Learning to Learn from Few Examples , 2019, ICLR.
[17] Daan Wierstra,et al. Meta-Learning with Memory-Augmented Neural Networks , 2016, ICML.
[18] Henry Zhu,et al. ROBEL: Robotics Benchmarks for Learning with Low-Cost Robots , 2019, CoRL.
[19] Ilya Kostrikov,et al. Image Augmentation Is All You Need: Regularizing Deep Reinforcement Learning from Pixels , 2020, ArXiv.
[20] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[21] Ladislau Bölöni,et al. Unsupervised Meta-Learning for Few-Shot Image Classification , 2019, NeurIPS.
[22] Christof Monz,et al. Data Augmentation for Low-Resource Neural Machine Translation , 2017, ACL.
[23] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[24] Yoshua Bengio,et al. Learning a synaptic learning rule , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.
[25] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[26] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[27] Samy Bengio,et al. Rapid Learning or Feature Reuse? Towards Understanding the Effectiveness of MAML , 2020, ICLR.
[28] Jascha Sohl-Dickstein,et al. Sensitivity and Generalization in Neural Networks: an Empirical Study , 2018, ICLR.
[29] Richard S. Zemel,et al. Prototypical Networks for Few-shot Learning , 2017, NIPS.
[30] Sanjeev Khudanpur,et al. Audio augmentation for speech recognition , 2015, INTERSPEECH.
[31] Mubarak Shah,et al. Task Agnostic Meta-Learning for Few-Shot Learning , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[32] Sebastian Thrun,et al. Lifelong Learning Algorithms , 1998, Learning to Learn.
[33] Katja Hofmann,et al. Fast Context Adaptation via Meta-Learning , 2018, ICML.
[34] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[35] Jascha Sohl-Dickstein,et al. Meta-Learning Update Rules for Unsupervised Representation Learning , 2018, ICLR.
[36] Jieyu Zhao,et al. Simple Principles of Metalearning , 1996 .
[37] Amos J. Storkey,et al. How to train your MAML , 2018, ICLR.
[38] Amos J. Storkey,et al. Assume, Augment and Learn: Unsupervised Few-Shot Meta-Learning via Random Labels and Data Augmentation , 2019, ArXiv.
[39] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.