Deep online classification using pseudo-generative models
暂无分享,去创建一个
Marin Ferecatu | Michel Crucianu | Pierre Blanchart | Andrey Besedin | M. Crucianu | Marin Ferecatu | P. Blanchart | A. Besedin
[1] Suresh Venkatasubramanian,et al. Streamed Learning: One-Pass SVMs , 2009, IJCAI.
[2] Geoff Hulten,et al. Mining high-speed data streams , 2000, KDD '00.
[3] Larry P. Heck,et al. Class-incremental Learning via Deep Model Consolidation , 2019, 2020 IEEE Winter Conference on Applications of Computer Vision (WACV).
[4] Razvan Pascanu,et al. Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.
[5] Ross B. Girshick,et al. Mask R-CNN , 2017, 1703.06870.
[6] Michael McCloskey,et al. Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem , 1989 .
[7] Jun Yu,et al. Hierarchical Deep Click Feature Prediction for Fine-Grained Image Recognition , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[8] Honglak Lee,et al. Online Incremental Feature Learning with Denoising Autoencoders , 2012, AISTATS.
[9] Jitendra Malik,et al. Region-Based Convolutional Networks for Accurate Object Detection and Segmentation , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[10] Geoffrey I. Webb,et al. Characterizing concept drift , 2015, Data Mining and Knowledge Discovery.
[11] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[12] Rishi Sharma,et al. A Note on the Inception Score , 2018, ArXiv.
[13] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[14] Wee Keong Ng,et al. A survey on data stream clustering and classification , 2015, Knowledge and Information Systems.
[15] Ronald Kemker,et al. FearNet: Brain-Inspired Model for Incremental Learning , 2017, ICLR.
[16] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[17] Yan Liu,et al. Deep Generative Dual Memory Network for Continual Learning , 2017, ArXiv.
[18] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[19] Jason Weston,et al. Natural Language Processing (Almost) from Scratch , 2011, J. Mach. Learn. Res..
[20] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[21] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[22] Simon Osindero,et al. Conditional Generative Adversarial Nets , 2014, ArXiv.
[23] Derek Hoiem,et al. Learning without Forgetting , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[24] Chrisantha Fernando,et al. PathNet: Evolution Channels Gradient Descent in Super Neural Networks , 2017, ArXiv.
[25] Ira Assent,et al. Indexing density models for incremental learning and anytime classification on data streams , 2009, EDBT '09.
[26] Jiwon Kim,et al. Continual Learning with Deep Generative Replay , 2017, NIPS.
[27] Wojciech Zaremba,et al. Improved Techniques for Training GANs , 2016, NIPS.
[28] Tapani Raiko,et al. Learning Deep Belief Networks from Non-stationary Streams , 2012, ICANN.
[29] Léon Bottou,et al. Wasserstein GAN , 2017, ArXiv.
[30] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[31] Jianping Fan,et al. iPrivacy: Image Privacy Protection by Identifying Sensitive Objects via Deep Multi-Task Learning , 2017, IEEE Transactions on Information Forensics and Security.
[32] Jonathon Shlens,et al. Conditional Image Synthesis with Auxiliary Classifier GANs , 2016, ICML.
[33] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[34] Stefan Wermter,et al. Continual Lifelong Learning with Neural Networks: A Review , 2019, Neural Networks.
[35] Anthony V. Robins,et al. Catastrophic Forgetting, Rehearsal and Pseudorehearsal , 1995, Connect. Sci..
[36] Surya Ganguli,et al. Continual Learning Through Synaptic Intelligence , 2017, ICML.