Adaptive online incremental learning for evolving data streams

Abstract Recent years have witnessed growing interests in online incremental learning. However, there are three major challenges in this area. The first major difficulty is concept drift, that is, the probability distribution in the streaming data would change as the data arrives. The second major difficulty is catastrophic forgetting, that is, forgetting what we have learned before when learning new knowledge. The last one we often ignore is the learning of the latent representation. Only good latent representation can improve the prediction accuracy of the model. Our research builds on this observation and attempts to overcome these difficulties. To this end, we propose an Adaptive Online Incremental Learning for evolving data streams (AOIL). We use auto-encoder with the memory module, on the one hand, we obtained the latent features of the input, on the other hand, according to the reconstruction loss of the auto-encoder with memory module, we could successfully detect the existence of concept drift and trigger the update mechanism, adjust the model parameters in time. In addition, we divide features, which are derived from the activation of the hidden layers, into two parts, which are used to extract the common and private features respectively. By means of this approach, the model could learn the private features of the new coming instances, but do not forget what we have learned in the past (shared features), which reduces the occurrence of catastrophic forgetting. At the same time, to get the fusion feature vector we use the self-attention mechanism to effectively fuse the extracted features, which further improved the latent representation learning. Moreover, in order to further improve the robustness of the algorithm, we add the de-noising auto-encoder to original framework. Finally, we conduct extensive experiments on different datasets, and show that the proposed AOIL gets promising results and outperforms other state-of-the-art methods.

[1]  Gregory Ditzler,et al.  Incremental Learning of Concept Drift from Streaming Imbalanced Data , 2013, IEEE Transactions on Knowledge and Data Engineering.

[2]  Marcin Budka,et al.  Towards cost-sensitive adaptation: When is it worth updating your predictive model? , 2015, Neurocomputing.

[3]  João Gama,et al.  On evaluating stream learning algorithms , 2012, Machine Learning.

[4]  Edwin Lughofer,et al.  DEVDAN: Deep Evolving Denoising Autoencoder , 2020, Neurocomputing.

[5]  Honglak Lee,et al.  Online Incremental Feature Learning with Denoising Autoencoders , 2012, AISTATS.

[6]  Yi Li,et al.  The Relaxed Online Maximum Margin Algorithm , 1999, Machine Learning.

[7]  Yun Fu,et al.  Feature Selection Guided Auto-Encoder , 2017, AAAI.

[8]  William Nick Street,et al.  A streaming ensemble algorithm (SEA) for large-scale classification , 2001, KDD '01.

[9]  Thomas de Quincey [C] , 2000, The Works of Thomas De Quincey, Vol. 1: Writings, 1799–1820.

[10]  Tsuyoshi Murata,et al.  {m , 1934, ACML.

[11]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Albert Bifet,et al.  Spiking Neural Networks and Online Learning: An Overview and Perspectives , 2019, Neural Networks.

[13]  Byoung-Tak Zhang,et al.  Dual-Memory Deep Learning Architectures for Lifelong Learning of Everyday Human Behaviors , 2016, IJCAI.

[14]  Alexandra-Bianca Borlea,et al.  Evolving Fuzzy Models for Prosthetic Hand Myoelectric-Based Control , 2020, IEEE Transactions on Instrumentation and Measurement.

[15]  Salvatore J. Stolfo,et al.  Cost-based modeling for fraud and intrusion detection: results from the JAM project , 2000, Proceedings DARPA Information Survivability Conference and Exposition. DISCEX'00.

[16]  Pavel Brazdil,et al.  Metalearning and Algorithm Selection: progress, state of the art and introduction to the 2018 Special Issue , 2017, Machine Learning.

[17]  Qian Du,et al.  Anomaly Detection and Reconstruction From Random Projections , 2012, IEEE Transactions on Image Processing.

[18]  Witold Pedrycz,et al.  An Incremental Construction of Deep Neuro Fuzzy System for Continual Learning of Nonstationary Data Streams , 2018, IEEE Transactions on Fuzzy Systems.

[19]  Koby Crammer,et al.  New Adaptive Algorithms for Online Classification , 2010, NIPS.

[20]  Koby Crammer,et al.  Learning via Gaussian Herding , 2010, NIPS.

[21]  Pascal Vincent,et al.  Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[22]  P. Baldi,et al.  Searching for exotic particles in high-energy physics with deep learning , 2014, Nature Communications.

[23]  Max Mühlhäuser,et al.  Analyzing business process anomalies using autoencoders , 2018, Machine Learning.

[24]  Hongzhi Wang,et al.  Life-long learning based on dynamic combination model , 2017, Appl. Soft Comput..

[25]  Jerzy Stefanowski,et al.  Combining block-based and online methods in learning ensembles from concept drifting data streams , 2014, Inf. Sci..

[26]  Bartosz Krawczyk,et al.  Online ensemble learning with abstaining classifiers for drifting and noisy data streams , 2017, Appl. Soft Comput..

[27]  Ricard Gavaldà,et al.  Learning from Time-Changing Data with Adaptive Windowing , 2007, SDM.

[28]  Claudio Gentile,et al.  A Second-Order Perceptron Algorithm , 2002, SIAM J. Comput..

[29]  Steven C. H. Hoi,et al.  Exact Soft Confidence-Weighted Learning , 2012, ICML.

[30]  João Gama,et al.  A survey on concept drift adaptation , 2014, ACM Comput. Surv..

[31]  Bowen Zhou,et al.  A Structured Self-attentive Sentence Embedding , 2017, ICLR.

[32]  Albert Bifet Classifier Concept Drift Detection and the Illusion of Progress , 2017, ICAISC.

[33]  Magdalena Deckert Incremental Rule-Based Learners for Handling Concept Drift: An Overview , 2013 .

[34]  João Gama,et al.  Learning with Drift Detection , 2004, SBIA.

[35]  Houxiang Zhang,et al.  Online Fault Detection in Autonomous Ferries: Using Fault-Type Independent Spectral Anomaly Detection , 2020, IEEE Transactions on Instrumentation and Measurement.

[36]  Koby Crammer,et al.  Adaptive regularization of weight vectors , 2009, Machine Learning.

[37]  Edwin Lughofer,et al.  ATL: Autonomous Knowledge Transfer from Many Streaming Processes , 2019, CIKM.

[38]  P. Alam ‘A’ , 2021, Composites Engineering: An A–Z Guide.

[39]  Sung Ju Hwang,et al.  Lifelong Learning with Dynamically Expandable Networks , 2017, ICLR.

[40]  Herna L. Viktor,et al.  Fast Hoeffding Drift Detection Method for Evolving Data Streams , 2016, ECML/PKDD.

[41]  Matthieu Guillaumin,et al.  Incremental Learning of Random Forests for Large-Scale Image Classification , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[42]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[43]  Mahardhika Pratama,et al.  Autonomous Deep Learning: Continual Learning Approach for Dynamic Environments , 2018, SDM.

[44]  Steven C. H. Hoi,et al.  Online Deep Learning: Learning Deep Neural Networks on the Fly , 2017, IJCAI.

[45]  Martin Zinkevich,et al.  Online Convex Programming and Generalized Infinitesimal Gradient Ascent , 2003, ICML.

[46]  Razvan Pascanu,et al.  Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.

[47]  Seung-Hwan Bae,et al.  Confidence-Based Data Association and Discriminative Deep Appearance Learning for Robust Online Multi-Object Tracking , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[48]  Yun Sing Koh,et al.  Detecting concept change in dynamic data streams , 2013, Machine Learning.

[49]  Joaquim F. Martins-Filho,et al.  An evolutionary approach with surrogate models and network science concepts to design optical networks , 2015, Eng. Appl. Artif. Intell..

[50]  Robi Polikar,et al.  Incremental Learning of Concept Drift in Nonstationary Environments , 2011, IEEE Transactions on Neural Networks.

[51]  Koby Crammer,et al.  Exact Convex Confidence-Weighted Learning , 2008, NIPS.

[52]  Francesco Orabona,et al.  Online Learning Algorithms , 2021 .

[53]  Claudio Gentile,et al.  A New Approximate Maximal Margin Classification Algorithm , 2002, J. Mach. Learn. Res..

[54]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[55]  Luis M. Candanedo,et al.  Accurate occupancy detection of an office room from light, temperature, humidity and CO2 measurements using statistical learning models , 2016 .

[56]  Svetha Venkatesh,et al.  Memorizing Normality to Detect Anomaly: Memory-Augmented Deep Autoencoder for Unsupervised Anomaly Detection , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[57]  Patrick Siarry,et al.  A postural information based biometric authentification system employing S-transform, radial basis network and Kalman filtering , 2010 .