LLEDA - Lifelong Self-Supervised Domain Adaptation

Humans and animals have the ability to continuously learn new information over their lifetime without losing previously acquired knowledge. However, artificial neural networks struggle with this due to new information conflicting with old knowledge, resulting in catastrophic forgetting. The complementary learning systems (CLS) theory suggests that the interplay between hippocampus and neocortex systems enables long-term and efficient learning in the mammalian brain, with memory replay facilitating the interaction between these two systems to reduce forgetting. The proposed Lifelong Self-Supervised Domain Adaptation (LLEDA) framework draws inspiration from the CLS theory and mimics the interaction between two networks: a DA network inspired by the hippocampus that quickly adjusts to changes in data distribution and an SSL network inspired by the neocortex that gradually learns domain-agnostic general representations. LLEDA's latent replay technique facilitates communication between these two networks by reactivating and replaying the past memory latent representations to stabilise long-term generalisation and retention without interfering with the previously learned information. Extensive experiments demonstrate that the proposed method outperforms several other methods resulting in a long-term adaptation while being less prone to catastrophic forgetting when transferred to new domains.

[1]  G. Leontidis,et al.  Semantic Positive Pairs for Enhancing Contrastive Instance Discrimination , 2023, ArXiv.

[2]  A. Durrant,et al.  S-JEA: Stacked Joint Embedding Architectures for Self-Supervised Visual Representation Learning , 2023, ArXiv.

[3]  A. Durrant,et al.  HMSN: Hyperbolic Self-Supervised Learning by Clustering with Ideal Prototypes , 2023, ArXiv.

[4]  Xiang Li,et al.  Two-phase self-supervised pretraining for object re-identification , 2022, Knowl. Based Syst..

[5]  Kiran R. Amin,et al.  A Cross-Domain Semantic Similarity Measure and Multi-Source Domain Adaptation in Sentiment Analysis , 2022, 2022 International Conference on Augmented Intelligence and Sustainable Systems (ICAISS).

[6]  G. Leontidis,et al.  Premonition Net, A Multi-Timeline Transformer Network Architecture Towards Strawberry Tabletop Yield Forecasting , 2022, Comput. Electron. Agric..

[7]  Fevziye Irem Eyiokur,et al.  A survey on computer vision based human analysis in the COVID-19 era , 2022, Image and Vision Computing.

[8]  Fabio De Sousa Ribeiro,et al.  Learning with Capsules: A Survey , 2022, ArXiv.

[9]  Wei Lu,et al.  Graph convolutional networks in language and vision: A survey , 2022, Knowl. Based Syst..

[10]  Pengjiang Qian,et al.  Multi-Modality Fusion & Inductive Knowledge Transfer Underlying Non-Sparse Multi-Kernel Learning and Distribution Adaption , 2022, IEEE/ACM Transactions on Computational Biology and Bioinformatics.

[11]  Steven Hoi,et al.  Continual Learning, Fast and Slow , 2021, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[12]  Xianglin Miao,et al.  Interpretable visual reasoning: A survey , 2021, Image Vis. Comput..

[13]  Yann LeCun,et al.  VICReg: Variance-Invariance-Covariance Regularization for Self-Supervised Learning , 2021, ICLR.

[14]  A. Durrant,et al.  Hyperspherically regularized networks for self-supervision , 2021, Image Vis. Comput..

[15]  Tyler L. Hayes,et al.  Replay in Deep Learning: Current Approaches and Missing Biological Elements , 2021, Neural Computation.

[16]  Mamatha Thota,et al.  Contrastive Domain Adaptation , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[17]  Yann LeCun,et al.  Barlow Twins: Self-Supervised Learning via Redundancy Reduction , 2021, ICML.

[18]  David May,et al.  How might technology rise to the challenge of data sharing in agri-food? , 2021, Global Food Security.

[19]  Dongrui Wu,et al.  EEG-Based Driver Drowsiness Estimation Using an Online Multi-View and Transfer TSK Fuzzy System , 2021, IEEE Transactions on Intelligent Transportation Systems.

[20]  Sinisa Segvic,et al.  Dense open-set recognition based on training with noisy negative images , 2021, Image Vis. Comput..

[21]  Riccardo Volpi,et al.  Continual Adaptation of Visual Representations via Domain Randomization and Meta-learning , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Markus Reischl,et al.  Cuepervision: self-supervised learning for continuous domain adaptation without catastrophic forgetting , 2020, Image Vis. Comput..

[23]  Mark Swainson,et al.  Multi-source domain adaptation for quality control in retail food packaging , 2020, Comput. Ind..

[24]  Xinlei Chen,et al.  Exploring Simple Siamese Representation Learning , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[25]  Mark Swainson,et al.  A novel unified deep neural networks methodology for use by date recognition in retail food package image , 2020, Signal, Image and Video Processing.

[26]  Hava T. Siegelmann,et al.  Brain-inspired replay for continual learning with artificial neural networks , 2020, Nature Communications.

[27]  Giovanni Maria Farinella,et al.  An unsupervised domain adaptation scheme for single-stage artwork recognition in cultural sites , 2020, Image Vis. Comput..

[28]  Di Qiu,et al.  Gradient Regularized Contrastive Learning for Continual Domain Adaptation , 2020, AAAI.

[29]  Geoffrey E. Hinton,et al.  Big Self-Supervised Models are Strong Semi-Supervised Learners , 2020, NeurIPS.

[30]  Pierre H. Richemond,et al.  Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning , 2020, NeurIPS.

[31]  Geoffrey E. Hinton,et al.  A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.

[32]  Vincenzo Lomonaco,et al.  Latent Replay for Real-Time Continual Learning , 2019, 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[33]  Ross B. Girshick,et al.  Momentum Contrast for Unsupervised Visual Representation Learning , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[34]  Tyler L. Hayes,et al.  REMIND Your Neural Network to Prevent Catastrophic Forgetting , 2019, ECCV.

[35]  Chao Chen,et al.  Deep joint two-stream Wasserstein auto-encoder and selective attention alignment for unsupervised domain adaptation , 2019, Neural Computing and Applications.

[36]  Yandong Guo,et al.  Large Scale Incremental Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[37]  Chen-Yu Lee,et al.  Sliced Wasserstein Discrepancy for Unsupervised Domain Adaptation , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[38]  Marc'Aurelio Ranzato,et al.  On Tiny Episodic Memories in Continual Learning , 2019 .

[39]  G. Tesauro,et al.  Learning to Learn without Forgetting By Maximizing Transfer and Minimizing Interference , 2018, ICLR.

[40]  Bogdan Raducanu,et al.  Memory Replay GANs: learning to generate images from new categories without forgetting , 2018, NeurIPS.

[41]  Chao Chen,et al.  Joint Domain Alignment and Discriminative Feature Learning for Unsupervised Deep Domain Adaptation , 2018, AAAI.

[42]  Oriol Vinyals,et al.  Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.

[43]  Davide Maltoni,et al.  Continuous Learning in Single-Incremental-Task Scenarios , 2018, Neural Networks.

[44]  Stefan Wermter,et al.  Continual Lifelong Learning with Neural Networks: A Review , 2018, Neural Networks.

[45]  Trevor Darrell,et al.  Adapting to Continuously Shifting Domains , 2018, ICLR.

[46]  Alex Bewley,et al.  Incremental Adversarial Domain Adaptation for Continually Changing Environments , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[47]  Ronald Kemker,et al.  FearNet: Brain-Inspired Model for Incremental Learning , 2017, ICLR.

[48]  Yang You,et al.  Large Batch Training of Convolutional Networks , 2017, 1708.03888.

[49]  Sung Ju Hwang,et al.  Lifelong Learning with Dynamically Expandable Networks , 2017, ICLR.

[50]  Sethuraman Panchanathan,et al.  Deep Hashing Network for Unsupervised Domain Adaptation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[51]  Marc'Aurelio Ranzato,et al.  Gradient Episodic Memory for Continual Learning , 2017, NIPS.

[52]  Trevor Darrell,et al.  Adversarial Discriminative Domain Adaptation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[53]  Andrei A. Rusu,et al.  Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.

[54]  Christoph H. Lampert,et al.  iCaRL: Incremental Classifier and Representation Learning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[55]  Kate Saenko,et al.  Deep CORAL: Correlation Alignment for Deep Domain Adaptation , 2016, ECCV Workshops.

[56]  Junmo Kim,et al.  Less-forgetting Learning in Deep Neural Networks , 2016, ArXiv.

[57]  James L. McClelland,et al.  What Learning Systems do Intelligent Agents Need? Complementary Learning Systems Theory Updated , 2016, Trends in Cognitive Sciences.

[58]  Derek Hoiem,et al.  Learning without Forgetting , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[59]  Michael I. Jordan,et al.  Deep Transfer Learning with Joint Adaptation Networks , 2016, ICML.

[60]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[61]  François Laviolette,et al.  Domain-Adversarial Training of Neural Networks , 2015, J. Mach. Learn. Res..

[62]  Michael I. Jordan,et al.  Learning Transferable Features with Deep Adaptation Networks , 2015, ICML.

[63]  Trevor Darrell,et al.  Deep Domain Confusion: Maximizing for Domain Invariance , 2014, CVPR 2014.

[64]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[65]  Yuan Shi,et al.  Geodesic flow kernel for unsupervised domain adaptation , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[66]  Bernhard Schölkopf,et al.  A Kernel Two-Sample Test , 2012, J. Mach. Learn. Res..

[67]  Trevor Darrell,et al.  Adapting Visual Category Models to New Domains , 2010, ECCV.

[68]  J. O’Neill,et al.  Play it again: reactivation of waking experience and memory , 2010, Trends in Neurosciences.

[69]  James L. McClelland,et al.  Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory. , 1995, Psychological review.

[70]  B. McNaughton,et al.  Reactivation of hippocampal ensemble memories during sleep. , 1994, Science.

[71]  Sendai,et al.  2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) , 2020 .

[72]  Andrew Y. Ng,et al.  Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .

[73]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[74]  Michael McCloskey,et al.  Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem , 1989 .

[75]  Isabelle Guyon,et al.  Neural Network Recognizer for Hand-Written Zip Code Digits , 1988, NIPS.

[76]  A. Krizhevsky ImageNet Classification with Deep Convolutional Neural Networks , 2022 .