Unsupervised Continual Learning in Streaming Environments

A deep clustering network is desired for data streams because of its aptitude in extracting natural features thus bypassing the laborious feature engineering step. While automatic construction of the deep networks in streaming environments remains an open issue, it is also hindered by the expensive labeling cost of data streams rendering the increasing demand for unsupervised approaches. This paper presents an unsupervised approach of deep clustering network construction on the fly via simultaneous deep learning and clustering termed Autonomous Deep Clustering Network (ADCN). It combines the feature extraction layer and autonomous fully connected layer in which both network width and depth are self-evolved from data streams based on the bias-variance decomposition of reconstruction loss. The self-clustering mechanism is performed in the deep embedding space of every fully connected layer while the final output is inferred via the summation of cluster prediction score. Further, a latent-based regularization is incorporated to resolve the catastrophic forgetting issue. A rigorous numerical study has shown that ADCN produces better performance compared to its counterparts while offering fully autonomous construction of ADCN structure in streaming environments with the absence of any labeled samples for model updates. To support the reproducible research initiative, codes, supplementary material, and raw results of ADCN are made available in https://tinyurl.com/AutonomousDCN.

[1]  Geoff Holmes,et al.  MOA: Massive Online Analysis , 2010, J. Mach. Learn. Res..

[2]  I-Cheng Yeh,et al.  The comparisons of data mining techniques for the predictive accuracy of probability of default of credit card clients , 2009, Expert Syst. Appl..

[3]  Mahardhika Pratama,et al.  A Self-Evolving Mutually-Operative Recurrent Network-based Model for Online Tool Condition Monitoring in Delay Scenario , 2020, KDD.

[4]  João Gama,et al.  Data Stream Classification Guided by Clustering on Nonstationary Environments and Extreme Verification Latency , 2015, SDM.

[5]  Jesús S. Aguilar-Ruiz,et al.  Knowledge discovery from data streams , 2009, Intell. Data Anal..

[6]  Charu C. Aggarwal,et al.  On the Surprising Behavior of Distance Metrics in High Dimensional Spaces , 2001, ICDT.

[7]  Shirin Enshaeifar,et al.  Continual Learning Using Bayesian Neural Networks , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[8]  R. French Catastrophic forgetting in connectionist networks , 1999, Trends in Cognitive Sciences.

[9]  Mahardhika Pratama,et al.  Autonomous Deep Learning: Continual Learning Approach for Dynamic Environments , 2018, SDM.

[10]  Steven C. H. Hoi,et al.  Online Deep Learning: Learning Deep Neural Networks on the Fly , 2017, IJCAI.

[11]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[12]  Derek Hoiem,et al.  Learning without Forgetting , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[13]  Narasimhan Sundararajan,et al.  A generalized growing and pruning RBF (GGAP-RBF) neural network for function approximation , 2005, IEEE Transactions on Neural Networks.

[14]  Sung Ju Hwang,et al.  Lifelong Learning with Dynamically Expandable Networks , 2017, ICLR.

[15]  Sreenatha G. Anavatti,et al.  PALM: An Incremental Construction of Hyperplanes for Data Stream Regression , 2018, IEEE Transactions on Fuzzy Systems.

[16]  Surya Ganguli,et al.  Continual Learning Through Synaptic Intelligence , 2017, ICML.

[17]  Edwin Lughofer,et al.  DEVDAN: Deep Evolving Denoising Autoencoder , 2020, Neurocomputing.

[18]  Huan Liu,et al.  Toward integrating feature selection algorithms for classification and clustering , 2005, IEEE Transactions on Knowledge and Data Engineering.

[19]  Gan Sun,et al.  Continual Multiview Task Learning via Deep Matrix Factorization , 2020, IEEE Transactions on Neural Networks and Learning Systems.

[20]  Mohamad Abdul Hady,et al.  Weakly Supervised Deep Learning Approach in Streaming Environments , 2019, 2019 IEEE International Conference on Big Data (Big Data).

[21]  William Nick Street,et al.  A streaming ensemble algorithm (SEA) for large-scale classification , 2001, KDD '01.

[22]  Robi Polikar,et al.  COMPOSE: A Semisupervised Learning Framework for Initially Labeled Nonstationary Streaming Data , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[23]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[24]  Jeff Johnson,et al.  Billion-Scale Similarity Search with GPUs , 2017, IEEE Transactions on Big Data.

[25]  Xuelong Li,et al.  Towards Compact ConvNets via Structure-Sparsity Regularized Filter Pruning , 2019, ArXiv.

[26]  Matthijs Douze,et al.  Deep Clustering for Unsupervised Learning of Visual Features , 2018, ECCV.

[27]  Roland Vollgraf,et al.  Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.

[28]  Weiping Ding,et al.  Automatic Construction of Multi-layer Perceptron Network from Streaming Examples , 2019, CIKM.

[29]  James Smith,et al.  Unsupervised Continual Learning and Self-Taught Associative Memory Hierarchies , 2019, ArXiv.

[30]  John F. Canny,et al.  T-SNE-CUDA: GPU-Accelerated T-SNE and its Applications to Modern Data , 2018, 2018 30th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD).

[31]  Yee Whye Teh,et al.  Progress & Compress: A scalable framework for continual learning , 2018, ICML.

[32]  Yoshua Bengio,et al.  Extracting and composing robust features with denoising autoencoders , 2008, ICML '08.

[33]  Shih-Fu Chang,et al.  Semi-supervised hashing for scalable image retrieval , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[34]  Marc'Aurelio Ranzato,et al.  Gradient Episodic Memory for Continual Learning , 2017, NIPS.

[35]  Shuicheng Yan,et al.  Dual Adversarial Autoencoders for Clustering , 2020, IEEE Transactions on Neural Networks and Learning Systems.

[36]  Witold Pedrycz,et al.  Online Tool Condition Monitoring Based on Parsimonious Ensemble+ , 2017, IEEE Transactions on Cybernetics.

[37]  Alex Lamb,et al.  Deep Learning for Classical Japanese Literature , 2018, ArXiv.

[38]  Honglak Lee,et al.  Online Incremental Feature Learning with Denoising Autoencoders , 2012, AISTATS.

[39]  Razvan Pascanu,et al.  Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.