暂无分享,去创建一个
Elisa Ricci | Julien Mairal | Xavier Alameda-Pineda | Karteek Alahari | Enrico Fini | Victor G. Turrisi da Costa
[1] Yann LeCun,et al. Barlow Twins: Self-Supervised Learning via Redundancy Reduction , 2021, ICML.
[2] Razvan Pascanu,et al. Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.
[3] Xinlei Chen,et al. Exploring Simple Siamese Representation Learning , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[4] E. Ricci,et al. Online Continual Learning under Extreme Memory Constraints , 2020, European Conference on Computer Vision.
[5] Yee Whye Teh,et al. Continual Unsupervised Representation Learning , 2019, NeurIPS.
[6] Kaiming He,et al. Momentum Contrast for Unsupervised Visual Representation Learning , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[7] Oriol Vinyals,et al. Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.
[8] Phillip Isola,et al. Contrastive Multiview Coding , 2019, ECCV.
[9] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[10] Cordelia Schmid,et al. End-to-End Incremental Learning , 2018, ECCV.
[11] Razvan Pascanu,et al. Progressive Neural Networks , 2016, ArXiv.
[12] Surya Ganguli,et al. Continual Learning Through Synaptic Intelligence , 2017, ICML.
[13] James Smith,et al. Unsupervised Progressive Learning and the STAM Architecture , 2019 .
[14] Marc'Aurelio Ranzato,et al. Gradient Episodic Memory for Continual Learning , 2017, NIPS.
[15] Philip H. S. Torr,et al. GDumb: A Simple Approach that Questions Our Progress in Continual Learning , 2020, ECCV.
[16] R. French. Catastrophic forgetting in connectionist networks , 1999, Trends in Cognitive Sciences.
[17] Nicu Sebe,et al. Solo-learn: A Library of Self-supervised Methods for Visual Representation Learning , 2021, ArXiv.
[18] Geoffrey E. Hinton,et al. A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.
[19] Kaiming He,et al. Improved Baselines with Momentum Contrastive Learning , 2020, ArXiv.
[20] Dahua Lin,et al. Learning a Unified Classifier Incrementally via Rebalancing , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[21] Yang You,et al. Large Batch Training of Convolutional Networks , 2017, 1708.03888.
[22] Jiwon Kim,et al. Continual Learning with Deep Generative Replay , 2017, NIPS.
[23] Yunxin Liu,et al. Rethinking the Representational Continuity: Towards Unsupervised Continual Learning , 2021, ArXiv.
[24] Jinwoo Shin,et al. Co2L: Contrastive Continual Learning , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[25] Ce Liu,et al. Supervised Contrastive Learning , 2020, NeurIPS.
[26] Matthias De Lange,et al. Continual learning: A comparative study on how to defy forgetting in classification tasks , 2019, ArXiv.
[27] Marcus Rohrbach,et al. Memory Aware Synapses: Learning what (not) to forget , 2017, ECCV.
[28] Stella X. Yu,et al. Unsupervised Feature Learning via Non-parametric Instance Discrimination , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[29] Joelle Pineau,et al. SPeCiaL: Self-Supervised Pretraining for Continual Learning , 2021, CSSL.
[30] Nicu Sebe,et al. Whitening for Self-Supervised Representation Learning , 2020, ICML.
[31] Matthijs Douze,et al. Deep Clustering for Unsupervised Learning of Visual Features , 2018, ECCV.
[32] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[33] Alexandros Karatzoglou,et al. Overcoming Catastrophic Forgetting with Hard Attention to the Task , 2018 .
[34] Matthieu Cord,et al. PODNet: Pooled Outputs Distillation for Small-Tasks Incremental Learning , 2020, ECCV.
[35] Aäron van den Oord,et al. Divide and Contrast: Self-supervised Learning from Uncurated Data , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[36] Yoshua Bengio,et al. An Empirical Investigation of Catastrophic Forgeting in Gradient-Based Neural Networks , 2013, ICLR.
[37] Jean Ponce,et al. VICReg: Variance-Invariance-Covariance Regularization for Self-Supervised Learning , 2021, ArXiv.
[38] Derek Hoiem,et al. Learning without Forgetting , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[39] Tyler L. Hayes,et al. Self-Supervised Training Enhances Online Continual Learning , 2021, BMVC.
[40] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[41] Anthony V. Robins,et al. Catastrophic Forgetting, Rehearsal and Pseudorehearsal , 1995, Connect. Sci..
[42] Yongtao Wang,et al. Continual Contrastive Self-supervised Learning for Image Classification , 2021, ArXiv.
[43] Philip H. S. Torr,et al. Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence , 2018, ECCV.
[44] Marc'Aurelio Ranzato,et al. Efficient Lifelong Learning with A-GEM , 2018, ICLR.
[45] Michael McCloskey,et al. Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem , 1989 .
[46] Cordelia Schmid,et al. Memory-Efficient Incremental Learning Through Feature Adaptation , 2020, ECCV.
[47] Julien Mairal,et al. Emerging Properties in Self-Supervised Vision Transformers , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[48] Julien Mairal,et al. Unsupervised Learning of Visual Features by Contrasting Cluster Assignments , 2020, NeurIPS.
[49] Christoph H. Lampert,et al. iCaRL: Incremental Classifier and Representation Learning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[50] Michal Valko,et al. Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning , 2020, NeurIPS.
[51] Kyunghyun Cho,et al. AAVAE: Augmentation-Augmented Variational Autoencoders , 2021, ArXiv.
[52] Simone Calderara,et al. Dark Experience for General Continual Learning: a Strong, Simple Baseline , 2020, NeurIPS.
[53] Yandong Guo,et al. Large Scale Incremental Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[54] Tom Eccles,et al. Life-Long Disentangled Representation Learning with Cross-Domain Latent Homologies , 2018, NeurIPS.
[55] Bo Wang,et al. Moment Matching for Multi-Source Domain Adaptation , 2018, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[56] Aapo Hyvärinen,et al. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models , 2010, AISTATS.
[57] Patrick Jähnichen,et al. Learning to Remember: A Synaptic Plasticity Driven Framework for Continual Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[58] Jonathan Tompson,et al. With a Little Help from My Friends: Nearest-Neighbor Contrastive Learning of Visual Representations , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).