Lessons learned from the NeurIPS 2021 MetaDL challenge: Backbone fine-tuning without episodic meta-learning dominates for few-shot learning image classification

Although deep neural networks are capable of achieving performance superior to humans on various tasks, they are notorious for requiring large amounts of data and computing resources, restricting their success to domains where such resources are available. Meta-learning methods can address this problem by transferring knowledge from related tasks, thus reducing the amount of data and computing resources needed to learn new tasks. We organize the MetaDL competition series, which provide opportunities for research groups all over the world to create and experimentally assess new meta-(deep)learning solutions for real problems. In this paper, authored collaboratively between the competition organizers and the top-ranked participants, we describe the design of the competition, the datasets, the best experimental results, as well as the top-ranked methods in the NeurIPS 2021 challenge , which attracted 15 active teams who made it to the final phase (by outper-forming the baseline), making over 100 code submissions during the feedback phase. The solutions of the top participants have been open-sourced . The lessons learned include that learning good representations is essential for effective transfer learning.

[1]  Timothy M. Hospedales,et al.  Pushing the Limits of Simple Pipelines for Few-Shot Learning: External Data and Fine-Tuning Make a Difference , 2022, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[2]  J. N. Rijn,et al.  Learning Curves for Decision Making in Supervised Machine Learning - A Survey , 2022, ArXiv.

[3]  Aske Plaat,et al.  Stateless neural meta-learning using second-order gradients , 2021, Machine Learning.

[4]  Timothy M. Hospedales,et al.  Meta-Learning in Neural Networks: A Survey , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  Jan N. van Rijn,et al.  Metalearning: Applications to Automated Machine Learning and Data Mining , 2022 .

[6]  J. Vanschoren,et al.  Advances in MetaDL: AAAI 2021 challenge and workshop , 2022, MetaDL@AAAI.

[7]  Isabelle Guyon,et al.  OmniPrint: A Configurable Printed Character Synthesizer , 2022, NeurIPS Datasets and Benchmarks.

[8]  Thomas M. Epalle,et al.  Aircraft Numerical “Twin”: A Time Series Regression Competition , 2021, International Conference on Machine Learning and Applications.

[9]  Massimiliano Pontil,et al.  The Role of Global Labels in Few-Shot Classification and How to Infer Them , 2021, NeurIPS.

[10]  Han Zhao,et al.  Bridging Multi-Task Learning and Meta-Learning: Towards Efficient Training and Effective Adaptation , 2021, ICML.

[11]  Julien Mairal,et al.  Emerging Properties in Self-Supervised Vision Transformers , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[12]  Frank Hutter,et al.  TrivialAugment: Tuning-free Yet State-of-the-Art Data Augmentation , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[13]  Amrith Rajagopal Setlur,et al.  Two Sides of Meta-Learning Evaluation: In vs. Out of Distribution , 2021, NeurIPS.

[14]  Xin Wang,et al.  MetaDelta: A Meta-Learning System for Few-shot Image Classification , 2021, MetaDL@AAAI.

[15]  Timothy M. Hospedales,et al.  Shallow Bayesian Meta Learning for Real-World Few-Shot Recognition , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[16]  Luca Bertinetto,et al.  On Episodes, Prototypical Networks, and Few-shot Learning , 2020, NeurIPS.

[17]  Timothy M. Hospedales,et al.  How Well Do Self-Supervised Models Transfer? , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Aske Plaat,et al.  A survey of deep meta-learning , 2020, Artificial Intelligence Review.

[19]  Sergio Escalera,et al.  Winning Solutions and Post-Challenge Analyses of the ChaLearn AutoDL Challenge 2019 , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[20]  Vincent Gripon,et al.  Leveraging the Feature Distribution in Transfer-based Few-Shot Learning , 2020, ICANN.

[21]  Isabelle Guyon,et al.  Bayesian Optimization is Superior to Random Search for Machine Learning Hyperparameter Tuning: Analysis of the Black-Box Optimization Challenge 2020 , 2021, NeurIPS.

[22]  Jiang Lu,et al.  Learning from Very Few Samples: A Survey , 2020, ArXiv.

[23]  Hugo Terashima-Marín,et al.  Learning from Few Samples: A Survey , 2020, ArXiv.

[24]  Yue Wang,et al.  Rethinking Few-Shot Image Classification: a Good Embedding Is All You Need? , 2020, ECCV.

[25]  Cordelia Schmid,et al.  Optimized Generic Feature Learning for Few-shot Classification across Domains , 2020, ArXiv.

[26]  Hugo Larochelle,et al.  Meta-Dataset: A Dataset of Datasets for Learning to Learn from Few Examples , 2019, ICLR.

[27]  Haifeng Li,et al.  RSI-CB: A Large Scale Remote Sensing Image Classification Benchmark via Crowdsource Data , 2017, ArXiv.

[28]  R. Julliard,et al.  Data Quality and Participant Engagement in Citizen Science: Comparing Two Approaches for Monitoring Pollinators in France and South Korea , 2019, Citizen Science: Theory and Practice.

[29]  R. GeethaRamani,et al.  Identification of plant leaf diseases using a nine-layer deep convolutional neural network , 2019, Comput. Electr. Eng..

[30]  James T. Kwok,et al.  Generalizing from a Few Examples , 2019, ACM Comput. Surv..

[31]  Yu-Chiang Frank Wang,et al.  A Closer Look at Few-shot Classification , 2019, ICLR.

[32]  Subhransu Maji,et al.  Task2Vec: Task Embedding for Meta-Learning , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[33]  Razvan Pascanu,et al.  Meta-Learning with Latent Embedding Optimization , 2018, ICLR.

[34]  Lars Kotthoff,et al.  The algorithm selection competitions 2015 and 2017 , 2018, Artif. Intell..

[35]  Sergio Escalera,et al.  Towards Automated Deep Learning: Analysis of the AutoDL challenge series 2019 , 2019, Proceedings of Machine Learning Research.

[36]  Joaquin Vanschoren,et al.  Meta-Learning: A Survey , 2018, Automated Machine Learning.

[37]  Chao Yang,et al.  A Survey on Deep Transfer Learning , 2018, ICANN.

[38]  Eyke Hüllermeier,et al.  ML-Plan: Automated machine learning via hierarchical planning , 2018, Machine Learning.

[39]  Joshua Achiam,et al.  On First-Order Meta-Learning Algorithms , 2018, ArXiv.

[40]  Joshua B. Tenenbaum,et al.  Meta-Learning for Semi-Supervised Few-Shot Classification , 2018, ICLR.

[41]  Tao Xiang,et al.  Learning to Compare: Relation Network for Few-Shot Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[42]  Pieter Abbeel,et al.  A Simple Neural Attentive Meta-Learner , 2017, ICLR.

[43]  Lars Kotthoff,et al.  Open Algorithm Selection Challenge 2017: Setup and Scenarios , 2017, OASC.

[44]  Hang Li,et al.  Meta-SGD: Learning to Learn Quickly for Few Shot Learning , 2017, ArXiv.

[45]  Richard S. Zemel,et al.  Prototypical Networks for Few-shot Learning , 2017, NIPS.

[46]  Sergey Levine,et al.  Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.

[47]  Hong Yu,et al.  Meta Networks , 2017, ICML.

[48]  Xiaoqiang Lu,et al.  Remote Sensing Image Scene Classification: Benchmark and State of the Art , 2017, Proceedings of the IEEE.

[49]  Hugo Larochelle,et al.  Optimization as a Model for Few-Shot Learning , 2016, ICLR.

[50]  Sergey Ioffe,et al.  Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.

[51]  Randal S. Olson,et al.  TPOT: A Tree-based Pipeline Optimization Tool for Automating Machine Learning , 2016, AutoML@ICML.

[52]  Kai Wang,et al.  A Benchmark for Automatic Visual Classification of Clinical Skin Disease Images , 2016, ECCV.

[53]  Daan Wierstra,et al.  Meta-Learning with Memory-Augmented Neural Networks , 2016, ICML.

[54]  Oriol Vinyals,et al.  Matching Networks for One Shot Learning , 2016, NIPS.

[55]  Aaron Klein,et al.  Efficient and Robust Automated Machine Learning , 2015, NIPS.

[56]  Marcel Salathé,et al.  An open access repository of images on plant health to enable the development of mobile disease diagnostics through machine learning and crowdsourcing , 2015, ArXiv.

[57]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[58]  Jian Sun,et al.  Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[59]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[60]  Marius Thomas Lindauer,et al.  AutoFolio: An Automatically Configured Algorithm Selector , 2015, J. Artif. Intell. Res..

[61]  Ivan Laptev,et al.  Learning and Transferring Mid-level Image Representations Using Convolutional Neural Networks , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[62]  Iasonas Kokkinos,et al.  Describing Textures in the Wild , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[63]  Yifan Gong,et al.  Cross-language knowledge transfer using multilingual deep neural network with shared hidden layers , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[64]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[65]  Pietro Perona,et al.  The Caltech-UCSD Birds-200-2011 Dataset , 2011 .

[66]  Gustaf Kylberg,et al.  Kylberg Texture Dataset v. 1.0 , 2011 .

[67]  Qiang Yang,et al.  A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.

[68]  Alireza Tavakoli Targhi,et al.  THE KTH-TIPS 2 database , 2006 .

[69]  Cordelia Schmid,et al.  A sparse texture representation using local affine regions , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[70]  Mario Fritz,et al.  THE KTH-TIPS database , 2004 .

[71]  Sebastian Thrun,et al.  Lifelong Learning Algorithms , 1998, Learning to Learn.

[72]  Richard J. Mammone,et al.  Meta-neural networks that learn by learning , 1992, [Proceedings 1992] IJCNN International Joint Conference on Neural Networks.