暂无分享,去创建一个
[1] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[2] Martha White,et al. Meta-Learning Representations for Continual Learning , 2019, NeurIPS.
[3] Quanming Yao,et al. Few-shot Learning: A Survey , 2019, ArXiv.
[4] Nicu Sebe,et al. Learning Deep Representations of Appearance and Motion for Anomalous Event Detection , 2015, BMVC.
[5] Philip H. S. Torr,et al. Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence , 2018, ECCV.
[6] M. M. Moya,et al. One-class classifier networks for target recognition applications , 1993 .
[7] Bernhard Schölkopf,et al. Estimating the Support of a High-Dimensional Distribution , 2001, Neural Computation.
[8] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[9] Tinne Tuytelaars,et al. Task-Free Continual Learning , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[11] Tom Schaul,et al. Prioritized Experience Replay , 2015, ICLR.
[12] Samy Bengio,et al. Rapid Learning or Feature Reuse? Towards Understanding the Effectiveness of MAML , 2020, ICLR.
[13] Gerald Tesauro,et al. Learning to Learn without Forgetting By Maximizing Transfer and Minimizing Interference , 2018, ICLR.
[14] Razvan Pascanu,et al. Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.
[15] Subhransu Maji,et al. Meta-Learning With Differentiable Convex Optimization , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[16] Zhiyuan Xu,et al. Adversarial Meta-Learning , 2018, ArXiv.
[17] Georg Langs,et al. Unsupervised Anomaly Detection with Generative Adversarial Networks to Guide Marker Discovery , 2017, IPMI.
[18] Rich Caruana,et al. Multitask Learning , 1998, Encyclopedia of Machine Learning and Data Mining.
[19] Geoffrey E. Hinton,et al. Reducing the Dimensionality of Data with Neural Networks , 2006, Science.
[20] Sergey Levine,et al. Meta-Learning and Universality: Deep Representations and Gradient Descent can Approximate any Learning Algorithm , 2017, ICLR.
[21] Richard S. Zemel,et al. Prototypical Networks for Few-shot Learning , 2017, NIPS.
[22] Surya Ganguli,et al. Continual Learning Through Synaptic Intelligence , 2017, ICML.
[23] Hongxing He,et al. Outlier Detection Using Replicator Neural Networks , 2002, DaWaK.
[24] Jiashi Feng,et al. Variational Prototype Replays for Continual Learning , 2019 .
[25] Razvan Pascanu,et al. Progressive Neural Networks , 2016, ArXiv.
[26] Christopher Leckie,et al. High-dimensional and large-scale anomaly detection using a linear one-class SVM with deep learning , 2016, Pattern Recognit..
[27] Chandan Srivastava,et al. Support Vector Data Description , 2011 .
[28] Hang Li,et al. Meta-SGD: Learning to Learn Quickly for Few Shot Learning , 2017, ArXiv.
[29] J. Schulman,et al. Reptile: a Scalable Metalearning Algorithm , 2018 .
[30] Mahmood Fathy,et al. Adversarially Learned One-Class Classifier for Novelty Detection , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[31] Sungzoon Cho,et al. Variational Autoencoder based Anomaly Detection using Reconstruction Probability , 2015 .
[32] Shehroz S. Khan,et al. One-class classification: taxonomy of study and review of techniques , 2013, The Knowledge Engineering Review.
[33] Marc'Aurelio Ranzato,et al. Gradient Episodic Memory for Continual Learning , 2017, NIPS.
[34] Hugo Larochelle,et al. Optimization as a Model for Few-Shot Learning , 2016, ICLR.
[35] VARUN CHANDOLA,et al. Anomaly detection: A survey , 2009, CSUR.
[36] Giacomo Spigler. Meta-learnt priors slow down catastrophic forgetting in neural networks , 2019, ArXiv.
[37] Volker Tresp,et al. Few-Shot One-Class Classification via Meta-Learning , 2019, AAAI.
[38] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[39] Luca Bertinetto,et al. Meta-learning with differentiable closed-form solvers , 2018, ICLR.
[40] Alexander Binder,et al. Deep One-Class Classification , 2018, ICML.
[41] Lewis D. Griffin,et al. Transfer representation-learning for anomaly detection , 2016, ICML 2016.
[42] R. French. Catastrophic forgetting in connectionist networks , 1999, Trends in Cognitive Sciences.
[43] Xin Yao,et al. A Systematic Study of Online Class Imbalance Learning With Concept Drift , 2017, IEEE Transactions on Neural Networks and Learning Systems.
[44] Joshua B. Tenenbaum,et al. One shot learning of simple visual concepts , 2011, CogSci.
[45] Yee Whye Teh,et al. Task Agnostic Continual Learning via Meta Learning , 2019, ArXiv.
[46] Nicu Sebe,et al. Abnormal event detection in videos using generative adversarial nets , 2017, 2017 IEEE International Conference on Image Processing (ICIP).
[47] Byoung-Tak Zhang,et al. Overcoming Catastrophic Forgetting by Incremental Moment Matching , 2017, NIPS.
[48] Charu C. Aggarwal,et al. Outlier Detection with Autoencoder Ensembles , 2017, SDM.
[49] James T. Kwok,et al. Generalizing from a Few Examples , 2019, ACM Comput. Surv..
[50] Benjamin Frederick Goodrich,et al. Neuron Clustering for Mitigating Catastrophic Forgetting in Supervised and Reinforcement Learning , 2015 .
[51] Joel Lehman,et al. Learning to Continually Learn , 2020, ECAI.