Learning Objective-Specific Active Learning Strategies with Attentive Neural Processes

Pool-based active learning (AL) is a promising technology for increasing data-efficiency of machine learning models. However, surveys show that performance of recent AL methods is very sensitive to the choice of dataset and training setting, making them unsuitable for general application. In order to tackle this problem, the field Learning Active Learning (LAL) suggests to learn the active learning strategy itself, allowing it to adapt to the given setting. In this work, we propose a novel LAL method for classification that exploits symmetry and independence properties of the active learning problem with an Attentive Conditional Neural Process model. Our approach is based on learning from a myopic oracle, which gives our model the ability to adapt to non-standard objectives, such as those that do not equally weight the error on all data points. We experimentally verify that our Neural Process model outperforms a variety of baselines in these settings. Finally, our experiments show that our model exhibits a tendency towards improved stability to changing datasets. However, performance is sensitive to choice of classifier and more work is necessary to reduce the performance the gap with the myopic oracle and to improve scalability. We present our work as a proof-of-concept for LAL on nonstandard objectives and hope our analysis and modelling considerations inspire future LAL work.

[1]  P. Indyk,et al.  Targeted Supervised Contrastive Learning for Long-Tailed Recognition , 2021, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[2]  Joost R. van Amersfoort,et al.  Causal-BALD: Deep Bayesian Active Learning of Outcomes to Infer Treatment-Effects from Observational Data , 2021, NeurIPS.

[3]  Bogdan Raducanu,et al.  Class-Balanced Active Learning for Image Classification , 2021, 2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV).

[4]  Weijia Li,et al.  Influence Selection for Active Learning , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[5]  Wolfgang Lehner,et al.  ImitAL: Learning Active Learning Strategies from Synthetic Data , 2021, ArXiv.

[6]  Tom Rainforth,et al.  Test Distribution-Aware Active Learning: A Principled Approach Against Distribution Shift and Outliers , 2021, 2106.11719.

[7]  Tae-Kyun Kim,et al.  Visual Transformer for Task-aware Active Learning , 2021, ArXiv.

[8]  Hao Wang,et al.  Delving into Deep Imbalanced Regression , 2021, ICML.

[9]  A. Yuille,et al.  CReST: A Class-Rebalancing Self-Training Framework for Imbalanced Semi-Supervised Learning , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[10]  A. Kirsch PowerEvaluationBALD: Efficient Evaluation-Oriented Deep (Bayesian) Active Learning with Stochastic Acquisition Functions , 2021, ArXiv.

[11]  V. Lemaire,et al.  Learning Active Learning at the Crossroads? Evaluation and Discussion , 2020, IAL@PKDD/ECML.

[12]  Antoni B. Chan,et al.  ALdataset: a benchmark for pool-based active learning , 2020, ArXiv.

[13]  Zhihui Li,et al.  A Survey of Deep Active Learning , 2020, ACM Comput. Surv..

[14]  Tae-Kyun Kim,et al.  Sequential Graph Convolutional Network for Active Learning , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  Zhi Xu,et al.  Rethinking the Value of Labels for Improving Class-Imbalanced Learning , 2020, NeurIPS.

[16]  Sandeep Tata,et al.  Active Learning for Skewed Data Sets , 2020, ArXiv.

[17]  Byoungjip Kim,et al.  VaB-AL: Incorporating Class Imbalance and Difficulty with Variational Bayes for Active Learning , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Céline Hudelot,et al.  Active Learning for Imbalanced Datasets , 2020, 2020 IEEE Winter Conference on Applications of Computer Vision (WACV).

[19]  Changjian Shui,et al.  Deep Active Learning: Unified and Principled Method for Query and Training , 2019, AISTATS.

[20]  José Miguel Hernández-Lobato,et al.  Bayesian Batch Active Learning as Sparse Subset Approximation , 2019, NeurIPS.

[21]  Shai Shalev-Shwartz,et al.  Discriminative Active Learning , 2019, ArXiv.

[22]  Melih Kandemir,et al.  Deep Active Learning with Adaptive Acquisition , 2019, IJCAI.

[23]  Nima Anari,et al.  Batch Active Learning Using Determinantal Point Processes , 2019, ArXiv.

[24]  John Langford,et al.  Deep Batch Active Learning by Diverse, Uncertain Gradient Lower Bounds , 2019, ICLR.

[25]  Minjie Xu,et al.  Understanding Goal-Oriented Active Learning via Influence Functions , 2019, ArXiv.

[26]  In So Kweon,et al.  Learning Loss for Active Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[27]  Gustavo Carneiro,et al.  Bayesian Generative Active Deep Learning , 2019, ICML.

[28]  Stella X. Yu,et al.  Large-Scale Long-Tailed Recognition in an Open World , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[29]  Trevor Darrell,et al.  Variational Adversarial Active Learning , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[30]  Taghi M. Khoshgoftaar,et al.  Survey on deep learning with class imbalance , 2019, J. Big Data.

[31]  José María Armingol,et al.  Balancing People Re-Identification Data for Deep Parts Similarity Learning , 2019, Journal of Imaging Science and Technology.

[32]  Qi Xie,et al.  Meta-Weight-Net: Learning an Explicit Mapping For Sample Weighting , 2019, NeurIPS.

[33]  Yee Whye Teh,et al.  Attentive Neural Processes , 2019, ICLR.

[34]  Pascal Fua,et al.  Discovering General-Purpose Active Learning Strategies , 2018, ArXiv.

[35]  Radu Timofte,et al.  Adversarial Sampling for Active Learning , 2018, 2020 IEEE Winter Conference on Applications of Computer Vision (WACV).

[36]  Yee Whye Teh,et al.  Neural Processes , 2018, ArXiv.

[37]  Yee Whye Teh,et al.  Conditional Neural Processes , 2018, ICML.

[38]  Andreas Nürnberger,et al.  The Power of Ensembles for Active Learning in Image Classification , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[39]  Dustin Tran,et al.  Image Transformer , 2018, ICML.

[40]  Yang Wu,et al.  Meta-Learning Transferable Active Learning Policies by Deep Reinforcement Learning , 2018, ArXiv.

[41]  Hugo Larochelle,et al.  Meta-Learning for Batch Mode Active Learning , 2018, ICLR.

[42]  Roland Vollgraf,et al.  Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.

[43]  Silvio Savarese,et al.  Active Learning for Convolutional Neural Networks: A Core-Set Approach , 2017, ICLR.

[44]  Pascal Fua,et al.  Learning Active Learning from Data , 2017, NIPS.

[45]  Zoubin Ghahramani,et al.  Deep Bayesian Active Learning with Image Data , 2017, ICML.

[46]  Maria Eugenia Ramirez-Loaiza,et al.  Active learning: an empirical study of common baselines , 2017, Data Mining and Knowledge Discovery.

[47]  José Bento,et al.  Generative Adversarial Active Learning , 2017, ArXiv.

[48]  Xiao Zhang,et al.  Range Loss for Deep Face Recognition with Long-Tailed Training Data , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[49]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[50]  Hsuan-Tien Lin,et al.  Active Learning by Learning , 2015, AAAI.

[51]  Christian Igel,et al.  Active learning with support vector machines , 2014, WIREs Data Mining Knowl. Discov..

[52]  L. Deng,et al.  The MNIST Database of Handwritten Digit Images for Machine Learning Research [Best of the Web] , 2012, IEEE Signal Processing Magazine.

[53]  Bernt Schiele,et al.  RALF: A reinforced active learning formulation for object class recognition , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[54]  Zoubin Ghahramani,et al.  Bayesian Active Learning for Classification and Preference Learning , 2011, ArXiv.

[55]  Byoung-Tak Zhang,et al.  Ensemble Learning with Active Example Selection for Imbalanced Biomedical Data Classification , 2011, IEEE/ACM Transactions on Computational Biology and Bioinformatics.

[56]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[57]  C. Lee Giles,et al.  Learning on the border: active learning in imbalanced data classification , 2007, CIKM '07.

[58]  Ran El-Yaniv,et al.  Online Choice of Active Learning Algorithms , 2003, J. Mach. Learn. Res..

[59]  Osvaldo Simeone,et al.  BAMLD: Bayesian Active Meta-Learning by Disagreement , 2021, ArXiv.

[60]  Jae Oh Woo,et al.  BABA: Beta Approximation for Bayesian Active Learning , 2021, ArXiv.

[61]  Gholamreza Haffari,et al.  Learning How to Actively Learn: A Deep Imitation Learning Approach , 2018, ACL.

[62]  Andrew Y. Ng,et al.  Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .

[63]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[64]  Burr Settles,et al.  Active Learning Literature Survey , 2009 .

[65]  John C. Platt,et al.  Probabilistic Outputs for Support vector Machines and Comparisons to Regularized Likelihood Methods , 1999 .