Augmented domain agreement for adaptable Meta-Learner on Few-Shot classification

Meta-learning approaches are recently introduced to obtain a proficient model in the problem of few-shot learning. The existing state-of-the-art methods can resolve the training difficulty to achieve the model generalization with limited instances by incorporating several tasks containing various drawn classes. However, during the fast-adaptation stage, the characteristics of instances in meta-training and meta-test sets are assumed to be similar. In contrast, most meta-learning algorithms are implemented in challenging settings where those are drawn from different populations. This critical assumption can cause the model to exhibit degraded performance. We propose an Augmented Domain Agreement for Adaptable Meta-Learner (AD2AML), which augments the domain adaptation framework in meta-learning to overcome this problem. We minimize the latent representation divergence of the inputs drawn from different distributions to enhance the model at obtaining more general features. Therefore, the trained network can be more transferable at shifted domain conditions. Furthermore, we extend our main idea by augmenting the image reconstruction network with cross-consistency loss to encourage the shared network to extract a similar input representation. We demonstrate our proposed method’s effectiveness on the benchmark datasets of few-shot classification and few-shot domain adaptation problems. Our experiment shows that our proposed idea can improve generalization performance. Moreover, the extension with image reconstruction and cross-consistency loss can stabilize domain loss minimization during training.

[1]  De-Chuan Zhan,et al.  Few-shot learning with adaptively initialized task optimizer: a practical meta-learning approach , 2019, Machine Learning.

[2]  R. A. Leibler,et al.  On Information and Sufficiency , 1951 .

[3]  Sebastian Thrun,et al.  Learning to Learn: Introduction and Overview , 1998, Learning to Learn.

[4]  Tao Xiang,et al.  Learning to Compare: Relation Network for Few-Shot Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[5]  Donald A. Adjeroh,et al.  Unified Deep Supervised Domain Adaptation and Generalization , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[6]  Jiashi Feng,et al.  Transferable Meta Learning Across Domains , 2018, UAI.

[7]  Richard S. Zemel,et al.  Prototypical Networks for Few-shot Learning , 2017, NIPS.

[8]  Pietro Perona,et al.  One-shot learning of object categories , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[9]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[10]  Sebastian Thrun,et al.  Learning to Learn , 1998, Springer US.

[11]  Sethuraman Panchanathan,et al.  Deep Hashing Network for Unsupervised Domain Adaptation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Amos Storkey,et al.  Meta-Learning in Neural Networks: A Survey , 2020, IEEE transactions on pattern analysis and machine intelligence.

[13]  Michael I. Jordan,et al.  Learning Transferable Features with Deep Adaptation Networks , 2015, ICML.

[14]  Abien Fred Agarap Deep Learning using Rectified Linear Units (ReLU) , 2018, ArXiv.

[15]  Sergey Levine,et al.  Meta-Learning with Implicit Gradients , 2019, NeurIPS.

[16]  Trevor Darrell,et al.  Deep Domain Confusion: Maximizing for Domain Invariance , 2014, CVPR 2014.

[17]  Trevor Darrell,et al.  Adapting Visual Category Models to New Domains , 2010, ECCV.

[18]  Kimin Lee,et al.  Using Pre-Training Can Improve Model Robustness and Uncertainty , 2019, ICML.

[19]  Taesung Park,et al.  CyCADA: Cycle-Consistent Adversarial Domain Adaptation , 2017, ICML.

[20]  Abhishek Sinha,et al.  Charting the Right Manifold: Manifold Mixup for Few-shot Learning , 2019, 2020 IEEE Winter Conference on Applications of Computer Vision (WACV).

[21]  Swami Sankaranarayanan,et al.  MetaReg: Towards Domain Generalization using Meta-Regularization , 2018, NeurIPS.

[22]  Taghi M. Khoshgoftaar,et al.  Deep learning applications and challenges in big data analytics , 2015, Journal of Big Data.

[23]  Seungjin Choi,et al.  Gradient-Based Meta-Learning with Learned Layerwise Metric and Subspace , 2018, ICML.

[24]  Joseph J. Lim,et al.  Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation , 2019, NeurIPS.

[25]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[26]  François Laviolette,et al.  Domain-Adversarial Training of Neural Networks , 2015, J. Mach. Learn. Res..

[27]  Aske Plaat,et al.  A survey of deep meta-learning , 2020, Artificial Intelligence Review.

[28]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.

[29]  Pietro Perona,et al.  Caltech-UCSD Birds 200 , 2010 .

[30]  Sergey Levine,et al.  Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.

[31]  Joshua B. Tenenbaum,et al.  One shot learning of simple visual concepts , 2011, CogSci.

[32]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[33]  Quoc V. Le,et al.  Do Better ImageNet Models Transfer Better? , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[34]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[35]  Joshua Achiam,et al.  On First-Order Meta-Learning Algorithms , 2018, ArXiv.

[36]  Razvan Pascanu,et al.  Meta-Learning with Warped Gradient Descent , 2020, ICLR.

[37]  Wouter M. Kouw An introduction to domain adaptation and transfer learning , 2018, ArXiv.

[38]  Jonathan Baxter Theoretical Models of Learning to Learn , 2020 .

[39]  Gregory R. Koch,et al.  Siamese Neural Networks for One-Shot Image Recognition , 2015 .

[40]  Joshua B. Tenenbaum,et al.  Meta-Learning for Semi-Supervised Few-Shot Classification , 2018, ICLR.

[41]  Trevor Darrell,et al.  Adversarial Discriminative Domain Adaptation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[42]  Hang Li,et al.  Meta-SGD: Learning to Learn Quickly for Few Shot Learning , 2017, ArXiv.

[43]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[44]  Victor S. Lempitsky,et al.  Unsupervised Domain Adaptation by Backpropagation , 2014, ICML.

[45]  Amos J. Storkey,et al.  How to train your MAML , 2018, ICLR.

[46]  Thomas L. Griffiths,et al.  Recasting Gradient-Based Meta-Learning as Hierarchical Bayes , 2018, ICLR.