Dropout as an Implicit Gating Mechanism For Continual Learning
暂无分享,去创建一个
Hassan Ghasemzadeh | Mehrdad Farajtabar | Seyed-Iman Mirzadeh | Seyed Iman Mirzadeh | Mehrdad Farajtabar | H. Ghasemzadeh
[1] Mehrdad Farajtabar,et al. Orthogonal Gradient Descent for Continual Learning , 2019, AISTATS.
[2] Nitish Srivastava,et al. Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.
[3] Nitish Srivastava,et al. Improving Neural Networks with Dropout , 2013 .
[4] Philip H. S. Torr,et al. Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence , 2018, ECCV.
[5] Kibok Lee,et al. Overcoming Catastrophic Forgetting With Unlabeled Data in the Wild , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[6] Colin Wei,et al. The Implicit and Explicit Regularization Effects of Dropout , 2020, ICML.
[7] Hassan Ghasemzadeh,et al. ActiLabel: A Combinatorial Transfer Learning Framework for Activity Recognition , 2020, ArXiv.
[8] Christoph H. Lampert,et al. iCaRL: Incremental Classifier and Representation Learning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[9] Svetlana Lazebnik,et al. PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[10] Albert Gordo,et al. Using Hindsight to Anchor Past Knowledge in Continual Learning , 2019, AAAI.
[11] G. Buzsáki,et al. The log-dynamic brain: how skewed distributions affect network operations , 2014, Nature Reviews Neuroscience.
[12] Razvan Pascanu,et al. Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.
[13] Derek Hoiem,et al. Learning without Forgetting , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[14] Marc'Aurelio Ranzato,et al. Efficient Lifelong Learning with A-GEM , 2018, ICLR.
[15] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[16] Hassan Ghasemzadeh,et al. Improved Knowledge Distillation via Teacher Assistant: Bridging the Gap Between Student and Teacher , 2019, ArXiv.
[17] Pierre Baldi,et al. Understanding Dropout , 2013, NIPS.
[18] Nicolas Y. Masse,et al. Alleviating catastrophic forgetting using context-dependent gating and synaptic stabilization , 2018, Proceedings of the National Academy of Sciences.
[19] Hossein Mobahi,et al. Self-Distillation Amplifies Regularization in Hilbert Space , 2020, NeurIPS.
[20] Yoshua Bengio,et al. An Empirical Investigation of Catastrophic Forgeting in Gradient-Based Neural Networks , 2013, ICLR.
[21] Michael McCloskey,et al. Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem , 1989 .
[22] Martial Mermillod,et al. The stability-plasticity dilemma: investigating the continuum from catastrophic forgetting to age-limited learning effects , 2013, Front. Psychol..
[23] Philip M. Long,et al. Surprising properties of dropout in deep networks , 2017, COLT.
[24] Xu Jia,et al. Continual learning: A comparative study on how to defy forgetting in classification tasks , 2019, ArXiv.
[25] Surya Ganguli,et al. Continual Learning Through Synaptic Intelligence , 2017, ICML.
[26] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.