An on-line weighted ensemble of regressor models to handle concept drifts

Abstract Many estimation, prediction, and learning applications have a dynamic nature. One of the most important challenges in machine learning is dealing with concept changes. Underlying changes may make the model designed on old data, inconsistent with new data. Also, algorithms usually specialize in one type of change. Other challenge is reusing previously acquired information in scenarios where changes may recur. This strategy improves the learning accuracy and reduces the processing time. Unfortunately, most existing learning algorithms to deal with changes are adapted on a batch basis. This process usually requires a long time, and such data may not reflect the current state of the system. However, even the system is adapted on a sample basis, existing algorithms may adapt slowly to changes and cannot conciliate old and new information. This paper proposes an On-line Weighted Ensemble (OWE) of regressor models which is able to learn incrementally sample by sample in the presence of several types of changes and simultaneously retain old information in recurring scenarios. The key idea is to keep a moving window that slides when a new sample is available. The error of each model on the current window is determined using a boosting strategy that assigns small errors to the models that predict accurately the samples predicted poorly by the ensemble. To handle recurring and non-recurring changes, OWE uses a new assignment of models׳ weights that takes into account the models׳ errors on the past and current windows using a discounting factor that decreases or increases the contribution of old windows. In addition, OWE launches new models if the system׳s accuracy is decreasing, and it can exclude inaccurate models over time. Experiments with artificial and industrial data reveal that in most cases OWE outperforms other state-of-the-art concept drift approaches.

[1]  Greg P. Timms,et al.  Ensemble aggregation methods for relocating models of rare events , 2014, Eng. Appl. Artif. Intell..

[2]  Robi Polikar,et al.  Incremental learning in nonstationary environments with controlled forgetting , 2009, 2009 International Joint Conference on Neural Networks.

[3]  Haiqing Wang,et al.  Soft Chemical Analyzer Development Using Adaptive Least-Squares Support Vector Regression with Selective Pruning and Variable Moving Window Size , 2009 .

[4]  Luigi Fortuna,et al.  Soft Sensors for Monitoring and Control of Industrial Processes (Advances in Industrial Control) , 2006 .

[5]  J. Freidman,et al.  Multivariate adaptive regression splines , 1991 .

[6]  Pablo H. Ibargüengoytia,et al.  Viscosity virtual sensor to control combustion in fossil fuel power plants , 2013, Eng. Appl. Artif. Intell..

[7]  S. D. Jong SIMPLS: an alternative approach to partial least squares regression , 1993 .

[8]  Žliobait . e,et al.  Learning under Concept Drift: an Overview , 2010 .

[9]  Slobodan Vucetic,et al.  Tracking Concept Change with Incremental Boosting by Minimization of the Evolving Exponential Loss , 2011, ECML/PKDD.

[10]  Claus Weihs,et al.  Variable window adaptive Kernel Principal Component Analysis for nonlinear nonstationary process monitoring , 2011, Comput. Ind. Eng..

[11]  Yeong-Koo Yeo,et al.  A recursive PLS-based soft sensor for prediction of the melt index during grade change operations in HDPE plant , 2009 .

[12]  Carlos Henggeler Antunes,et al.  A genetic algorithm for designing neural network ensembles , 2012, GECCO '12.

[13]  Rajandrea Sethi,et al.  Artificial neural network simulation of hourly groundwater levels in a coastal aquifer system of the Venice lagoon , 2012, Eng. Appl. Artif. Intell..

[14]  Durga L. Shrestha,et al.  Experiments with AdaBoost.RT, an Improved Boosting Scheme for Regression , 2006, Neural Computation.

[15]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[16]  Jerzy Stefanowski,et al.  Reacting to Different Types of Concept Drift: The Accuracy Updated Ensemble Algorithm , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[17]  Kwok-wing Chau,et al.  Application of a PSO-based neural network in analysis of outcomes of construction claims , 2007 .

[18]  Robi Polikar Ensemble learning , 2009, Scholarpedia.

[19]  Ralf Klinkenberg Meta-Learning, Model Selection, and Example Selection in Machine Learning Domains with Concept Drift , 2005, LWA.

[20]  Nikunj C. Oza,et al.  Online Ensemble Learning , 2000, AAAI/IAAI.

[21]  Marcus A. Maloof,et al.  Using additive expert ensembles to cope with concept drift , 2005, ICML.

[22]  Jon Atli Benediktsson,et al.  Multiple Classifier Systems , 2015, Lecture Notes in Computer Science.

[23]  Gerhard Widmer,et al.  Learning in the presence of concept drift and hidden contexts , 2004, Machine Learning.

[24]  João Gama,et al.  Enhancing data stream predictions with reliability estimators and explanation , 2014, Eng. Appl. Artif. Intell..

[25]  Soon Keat Tan,et al.  Localized, Adaptive Recursive Partial Least Squares Regression for Dynamic System Modeling , 2012 .

[26]  Kyosuke Nishida,et al.  Adaptive Classifiers-Ensemble System for Tracking Concept Drift , 2007, 2007 International Conference on Machine Learning and Cybernetics.

[27]  Takashi Omori,et al.  ACE: Adaptive Classifiers-Ensemble System for Concept-Drifting Environments , 2005, Multiple Classifier Systems.

[28]  Meng Chao,et al.  Neural network ensembles based on copula methods and Distributed Multiobjective Central Force Optimization algorithm , 2014, Eng. Appl. Artif. Intell..

[29]  Rui Araújo,et al.  Mixture of partial least squares experts and application in prediction settings with multiple operating modes , 2014 .

[30]  P. Kitanidis,et al.  Effects of kinetic mass transfer and transient flow conditions on widening mixing zones in coastal aquifers , 2009 .

[31]  Gang Li,et al.  An Assessment of Combining Tourism Demand Forecasts over Different Time Horizons , 2008 .

[32]  Xin Yao,et al.  The Impact of Diversity on Online Ensemble Learning in the Presence of Concept Drift , 2010, IEEE Transactions on Knowledge and Data Engineering.

[33]  Carlo Zaniolo,et al.  Fast and Light Boosting for Adaptive Mining of Data Streams , 2004, PAKDD.

[34]  Jun Zhang,et al.  Multilayer Ensemble Pruning via Novel Multi-sub-swarm Particle Swarm Optimization , 2009, J. Univers. Comput. Sci..

[35]  Kwok-Wing Chau,et al.  A new image thresholding method based on Gaussian mixture model , 2008, Appl. Math. Comput..

[36]  Xiao Fan Wang,et al.  Soft sensing modeling based on support vector machine and Bayesian model selection , 2004, Comput. Chem. Eng..

[37]  Indre Zliobaite,et al.  Learning under Concept Drift: an Overview , 2010, ArXiv.

[38]  Bogdan Gabrys,et al.  Review of adaptation mechanisms for data-driven soft sensors , 2011, Comput. Chem. Eng..

[39]  Tao Xiong,et al.  A combined SVM and LDA approach for classification , 2005, Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005..

[40]  Robi Polikar,et al.  Incremental Learning of Concept Drift in Nonstationary Environments , 2011, IEEE Transactions on Neural Networks.

[41]  Manfred K. Warmuth,et al.  The weighted majority algorithm , 1989, 30th Annual Symposium on Foundations of Computer Science.

[42]  Carlos Henggeler Antunes,et al.  Comparison of a genetic algorithm and simulated annealing for automatic neural network ensemble development , 2013, Neurocomputing.

[43]  Eyke Hüllermeier,et al.  IBLStreams: a system for instance-based classification and regression on data streams , 2012, Evol. Syst..

[44]  Xin Yao,et al.  DDD: A New Ensemble Approach for Dealing with Concept Drift , 2012, IEEE Transactions on Knowledge and Data Engineering.

[45]  S. Qin Recursive PLS algorithms for adaptive data modeling , 1998 .

[46]  K. Chau,et al.  Predicting monthly streamflow using data‐driven models coupled with data‐preprocessing techniques , 2009 .

[47]  Chuntian Cheng,et al.  Long-Term Prediction of Discharges in Manwan Reservoir Using Artificial Neural Network Models , 2005, ISNN.

[48]  Harris Drucker,et al.  Improving Regressors using Boosting Techniques , 1997, ICML.

[49]  Shankar Vembu,et al.  Chemical gas sensor drift compensation using classifier ensembles , 2012 .

[50]  Bogdan Gabrys,et al.  Local learning‐based adaptive soft sensor for catalyst activation prediction , 2011 .