On robustness of generative representations against catastrophic forgetting

Catastrophic forgetting of previously learned knowledge while learning new tasks is a widely observed limitation of contemporary neural networks. Although many continual learning methods are proposed to mitigate this drawback, the main question remains unanswered: what is the root cause of catastrophic forgetting? In this work, we aim at answering this question by posing and validating a set of research hypotheses related to the specificity of representations built internally by neural models. More specifically, we design a set of empirical evaluations that compare the robustness of representations in discriminative and generative models against catastrophic forgetting. We observe that representations learned by discriminative models are more prone to catastrophic forgetting than their generative counterparts, which sheds new light on the advantages of developing generative models for continual learning. Finally, our work opens new research pathways and possibilities to adopt generative models in continual learning beyond mere replay mechanisms.

[1]  Tinne Tuytelaars,et al.  Rehearsal revealed: The limits and merits of revisiting samples in continual learning , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[2]  Philip H. S. Torr,et al.  GDumb: A Simple Approach that Questions Our Progress in Continual Learning , 2020, ECCV.

[3]  Max Welling,et al.  Auto-Encoding Variational Bayes , 2013, ICLR.

[4]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[5]  Tinne Tuytelaars,et al.  Online Continual Learning with Maximally Interfered Retrieval , 2019, ArXiv.

[6]  Kyunghyun Cho,et al.  Continual Learning via Neural Pruning , 2019, ArXiv.

[7]  David Rolnick,et al.  Experience Replay for Continual Learning , 2018, NeurIPS.

[8]  Svetlana Lazebnik,et al.  Piggyback: Adapting a Single Network to Multiple Tasks by Learning to Mask Weights , 2018, ECCV.

[9]  Razvan Pascanu,et al.  Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.

[10]  Ali Farhadi,et al.  Supermasks in Superposition , 2020, NeurIPS.

[11]  Jiwon Kim,et al.  Continual Learning with Deep Generative Replay , 2017, NIPS.

[12]  Svetlana Lazebnik,et al.  PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[13]  Ethan Dyer,et al.  Anatomy of Catastrophic Forgetting: Hidden Representations and Task Semantics , 2020, ICLR.

[14]  Kamil Deja,et al.  BinPlay: A Binary Latent Autoencoder for Generative Replay Continual Learning , 2021, 2021 International Joint Conference on Neural Networks (IJCNN).

[15]  Benjamin F. Grewe,et al.  Continual learning with hypernetworks , 2019, ICLR.

[16]  Tian Han,et al.  A Tale of Three Probabilistic Families: Discriminative, Descriptive and Generative Models , 2018, Quarterly of Applied Mathematics.

[17]  Stefan Wermter,et al.  Continual Lifelong Learning with Neural Networks: A Review , 2019, Neural Networks.

[18]  Zhanxing Zhu,et al.  Reinforced Continual Learning , 2018, NeurIPS.

[19]  Joelle Pineau,et al.  Online Learned Continual Compression with Adaptive Quantization Modules , 2019, ICML.

[20]  Geoffrey E. Hinton,et al.  Similarity of Neural Network Representations Revisited , 2019, ICML.

[21]  Michael C. Mozer,et al.  Sequential Mastery of Multiple Visual Tasks: Networks Naturally Learn to Learn and Forget to Forget , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Andreas S. Tolias,et al.  Generative replay with feedback connections as a general strategy for continual learning , 2018, ArXiv.

[23]  R. French Catastrophic forgetting in connectionist networks , 1999, Trends in Cognitive Sciences.

[24]  Nicolas Y. Masse,et al.  Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization , 2018, Proceedings of the National Academy of Sciences.

[25]  Surya Ganguli,et al.  Continual Learning Through Synaptic Intelligence , 2017, ICML.

[26]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.