Large-scale Pre-trained Models are Surprisingly Strong in Incremental Novel Class Discovery

Discovering novel concepts from unlabelled data and in a continuous manner is an important desideratum of lifelong learners. In the literature such problems have been partially addressed under very restricted settings, where either access to labelled data is provided for discovering novel concepts (e.g., NCD) or learning occurs for a limited number of incremental steps (e.g., class-iNCD). In this work we challenge the status quo and propose a more challenging and practical learning paradigm called MSc-iNCD, where learning occurs continuously and unsupervisedly, while exploiting the rich priors from large-scale pre-trained models. To this end, we propose simple baselines that are not only resilient under longer learning scenarios, but are surprisingly strong when compared with sophisticated state-of-the-art methods. We conduct extensive empirical evaluation on a multitude of benchmarks and show the effectiveness of our proposed baselines, which significantly raises the bar.

[1]  V. Lemaire,et al.  Novel Class Discovery: an Introduction and Key Concepts , 2023, ArXiv.

[2]  Jun Zhu,et al.  A Comprehensive Survey of Continual Learning: Theory, Method and Application , 2023, ArXiv.

[3]  S. Calderara,et al.  Class-Incremental Continual Learning Into the eXtended DER-Verse , 2022, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Bingchen Zhao,et al.  XCon: Learning with Experts for Fine-grained Category Discovery , 2022, BMVC.

[5]  K. J. Joseph,et al.  Novel Class Discovery without Forgetting , 2022, ECCV.

[6]  N. Sebe,et al.  Class-incremental Novel Class Discovery , 2022, ECCV.

[7]  Jiaping Yu,et al.  Divide and Conquer: Compositional Experts for Generalized Novel Class Discovery , 2022, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[8]  S. Calderara,et al.  Transfer without Forgetting , 2022, ECCV.

[9]  N. Vasconcelos,et al.  Class-Incremental Learning with Strong Pre-trained Models , 2022, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[10]  T. Tuytelaars,et al.  Residual Tuning: Toward Novel Category Discovery Without Labels , 2022, IEEE Transactions on Neural Networks and Learning Systems.

[11]  A. Vedaldi,et al.  Generalized Category Discovery , 2022, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Alahari Karteek,et al.  Self-Supervised Models are Continual Learners , 2021, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[13]  Sung Ju Hwang,et al.  Representational Continuity for Unsupervised Continual Learning , 2021, ICLR.

[14]  Jure Leskovec,et al.  Open-World Semi-Supervised Learning , 2021, ICLR.

[15]  Xiaojuan Qi,et al.  A Simple Parametric Classification Baseline for Generalized Category Discovery , 2022, ArXiv.

[16]  Elisa Ricci,et al.  A Unified Objective for Novel Class Discovery , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[17]  K. Han,et al.  Novel Visual Category Discovery with Dual Ranking Statistics and Mutual Knowledge Distillation , 2021, NeurIPS.

[18]  Nicu Sebe,et al.  Neighborhood Contrastive Learning for Novel Class Discovery , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[19]  Fahad Shahbaz Khan,et al.  Intriguing Properties of Vision Transformers , 2021, NeurIPS.

[20]  Julien Mairal,et al.  Emerging Properties in Self-Supervised Vision Transformers , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[21]  Kai Han,et al.  Joint Representation Learning and Novel Category Discovery on Single- and Multi-modal Data , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[22]  Ilya Sutskever,et al.  Learning Transferable Visual Models From Natural Language Supervision , 2021, ICML.

[23]  S. Gelly,et al.  An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale , 2020, ICLR.

[24]  Jianping Gou,et al.  Knowledge Distillation: A Survey , 2020, International Journal of Computer Vision.

[25]  Nicu Sebe,et al.  OpenMix: Reviving Known Knowledge for Discovering Novel Visual Categories in an Open World , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[26]  A. Linear-probe,et al.  Learning Transferable Visual Models From Natural Language Supervision , 2021 .

[27]  Julien Mairal,et al.  Unsupervised Learning of Visual Features by Contrasting Cluster Assignments , 2020, NeurIPS.

[28]  Simone Calderara,et al.  Dark Experience for General Continual Learning: a Strong, Simple Baseline , 2020, NeurIPS.

[29]  Andrew Zisserman,et al.  Automatically Discovering and Learning New Visual Categories with Ranking Statistics , 2020, ICLR.

[30]  Geoffrey E. Hinton,et al.  A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.

[31]  Andrew Zisserman,et al.  Learning to Discover Novel Visual Categories via Deep Transfer Clustering , 2019 .

[32]  Serge J. Belongie,et al.  The Herbarium Challenge 2019 Dataset , 2019, ArXiv.

[33]  Dahua Lin,et al.  Learning a Unified Classifier Incrementally via Rebalancing , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[34]  Yandong Guo,et al.  Large Scale Incremental Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[35]  Zsolt Kira,et al.  Multi-class Classification without Multi-class Labels , 2019, ICLR.

[36]  Zsolt Kira,et al.  Learning to cluster in order to Transfer across domains and tasks , 2017, ICLR.

[37]  Jianfeng Zhan,et al.  Cosine Normalization: Using Cosine Similarity Instead of Dot Product in Neural Networks , 2017, ICANN.

[38]  Derek Hoiem,et al.  Learning without Forgetting , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[39]  Lingfeng Wang,et al.  Deep Adaptive Image Clustering , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[40]  Cheng Deng,et al.  Deep Clustering via Joint Convolutional Autoencoder Embedding and Relative Entropy Minimization , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[41]  Razvan Pascanu,et al.  Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.

[42]  Christoph H. Lampert,et al.  iCaRL: Incremental Classifier and Representation Learning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[43]  Bo Yang,et al.  Towards K-means-friendly Spaces: Simultaneous Deep Learning and Clustering , 2016, ICML.

[44]  Dhruv Batra,et al.  Joint Unsupervised Learning of Deep Representations and Image Clusters , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[45]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[46]  Ali Farhadi,et al.  Unsupervised Deep Embedding for Clustering Analysis , 2015, ICML.

[47]  Ya Le,et al.  Tiny ImageNet Visual Recognition Challenge , 2015 .

[48]  Marco Cuturi,et al.  Sinkhorn Distances: Lightspeed Computation of Optimal Transport , 2013, NIPS.

[49]  Pietro Perona,et al.  Caltech-UCSD Birds 200 , 2010 .

[50]  Anil K. Jain Data clustering: 50 years beyond K-means , 2008, Pattern Recognit. Lett..

[51]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[52]  Sergei Vassilvitskii,et al.  k-means++: the advantages of careful seeding , 2007, SODA '07.

[53]  R. French Catastrophic forgetting in connectionist networks , 1999, Trends in Cognitive Sciences.