暂无分享,去创建一个
[1] Byoung-Tak Zhang,et al. Overcoming Catastrophic Forgetting by Incremental Moment Matching , 2017, NIPS.
[2] Razvan Pascanu,et al. Progressive Neural Networks , 2016, ArXiv.
[3] Mingli Song,et al. Student Becoming the Master: Knowledge Amalgamation for Joint Scene Parsing, Depth Estimation, and More , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[4] Martha White,et al. Meta-Learning Representations for Continual Learning , 2019, NeurIPS.
[5] R Ratcliff,et al. Connectionist models of recognition memory: constraints imposed by learning and forgetting functions. , 1990, Psychological review.
[6] Li Sun,et al. Amalgamating Knowledge towards Comprehensive Classification , 2018, AAAI.
[7] M E J Newman,et al. Finding and evaluating community structure in networks. , 2003, Physical review. E, Statistical, nonlinear, and soft matter physics.
[8] David M. Reif,et al. Profiling of the Tox21 10K compound library for agonists and antagonists of the estrogen receptor alpha signaling pathway , 2014, Scientific Reports.
[9] Zunlei Feng,et al. Factorizable Graph Convolutional Networks , 2020, NeurIPS.
[10] Yee Whye Teh,et al. Progress & Compress: A scalable framework for continual learning , 2018, ICML.
[11] Trevor Darrell,et al. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.
[12] Tianqi Chen,et al. Net2Net: Accelerating Learning via Knowledge Transfer , 2015, ICLR.
[13] Svetlana Lazebnik,et al. PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[14] Michael McCloskey,et al. Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem , 1989 .
[15] Yoshua Bengio,et al. Gradient based sample selection for online continual learning , 2019, NeurIPS.
[16] Marc'Aurelio Ranzato,et al. Gradient Episodic Memory for Continual Learning , 2017, NIPS.
[17] Xavier Bresson,et al. Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering , 2016, NIPS.
[18] Jiwon Kim,et al. Continual Learning with Deep Generative Replay , 2017, NIPS.
[19] Jure Leskovec,et al. Predicting multicellular function through multi-layer tissue networks , 2017, Bioinform..
[20] Stephan Günnemann,et al. Deep Gaussian Embedding of Graphs: Unsupervised Inductive Learning via Ranking , 2017, ICLR.
[21] R. French. Dynamically constraining connectionist networks to produce distributed, orthogonal representations to reduce catastrophic interference , 2019, Proceedings of the Sixteenth Annual Conference of the Cognitive Science Society.
[22] Jure Leskovec,et al. Inductive Representation Learning on Large Graphs , 2017, NIPS.
[23] Pietro Liò,et al. Graph Attention Networks , 2017, ICLR.
[24] Chrisantha Fernando,et al. PathNet: Evolution Channels Gradient Descent in Super Neural Networks , 2017, ArXiv.
[25] Anthony V. Robins,et al. Catastrophic Forgetting, Rehearsal and Pseudorehearsal , 1995, Connect. Sci..
[26] Reinhard Schneider,et al. Using graph theory to analyze biological networks , 2011, BioData Mining.
[27] Christoph H. Lampert,et al. iCaRL: Incremental Classifier and Representation Learning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[28] Anton van den Hengel,et al. Image-Based Recommendations on Styles and Substitutes , 2015, SIGIR.
[29] Gerald Tesauro,et al. Learning to Learn without Forgetting By Maximizing Transfer and Minimizing Interference , 2018, ICLR.
[30] Richard E. Turner,et al. Variational Continual Learning , 2017, ICLR.
[31] Rich Caruana,et al. Multitask Learning , 1998, Encyclopedia of Machine Learning and Data Mining.
[32] Pascal Fua,et al. Tracking Interacting Objects Optimally Using Integer Programming , 2014, ECCV.
[33] Sung Ju Hwang,et al. Lifelong Learning with Dynamically Expandable Networks , 2017, ICLR.
[34] D. Tao,et al. Distilling Knowledge From Graph Convolutional Networks , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[35] Surya Ganguli,et al. Continual Learning Through Synaptic Intelligence , 2017, ICML.
[36] Max Welling,et al. Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.
[37] Dacheng Tao,et al. Hallucinating Visual Instances in Total Absentia , 2020, European Conference on Computer Vision.
[38] Razvan Pascanu,et al. Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.
[39] King-Sun Fu,et al. IEEE Transactions on Pattern Analysis and Machine Intelligence Publication Information , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[40] Marcus Rohrbach,et al. Memory Aware Synapses: Learning what (not) to forget , 2017, ECCV.
[41] Alexandros Karatzoglou,et al. Overcoming Catastrophic Forgetting with Hard Attention to the Task , 2018 .
[42] Joan Bruna,et al. Spectral Networks and Locally Connected Networks on Graphs , 2013, ICLR.
[43] Marc'Aurelio Ranzato,et al. Efficient Lifelong Learning with A-GEM , 2018, ICLR.
[44] James L. McClelland,et al. Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory. , 1995, Psychological review.
[45] Dacheng Tao,et al. SPAGAN: Shortest Path Graph Attention Network , 2019, IJCAI.
[46] R. French. Catastrophic forgetting in connectionist networks , 1999, Trends in Cognitive Sciences.
[47] Jure Leskovec,et al. How Powerful are Graph Neural Networks? , 2018, ICLR.
[48] Derek Hoiem,et al. Learning without Forgetting , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.