Fast adaptive stacking of ensembles

This paper presents a new ensemble method for learning from non-stationary data streams. In these situations, massive data are constantly generated at high speed and their target function can change over time. The proposed method, named Fast Adaptive Stacking of Ensembles (FASE), uses a meta-classifier to combine the predictions from the base classifiers in the ensemble. FASE maintains a set of adaptive learners, in order to deal with concept drifting data. The new algorithm is able to process the input data in constant time and space computational complexity. It only receives as parameters the confidence level for the change detection mechanism and the number of base classifiers. These characteristics make FASE very suitable for learning from non-stationary data streams. We empirically compare the new algorithm with various state-of-the-art ensemble methods for learning in non-stationary data streams. We use a Naïve Bayes classifier and a Perceptron to evaluate the performance of the algorithms over real-world datasets. The experiment results show that FASE presents higher predictive accuracy in the investigated tasks, being also able to bound its computational cost.

[1]  Geoff Holmes,et al.  Ensembles of Restricted Hoeffding Trees , 2012, TIST.

[2]  João Gama,et al.  Evaluating algorithms that learn from data streams , 2009, SAC '09.

[3]  João Gama,et al.  On evaluating stream learning algorithms , 2012, Machine Learning.

[4]  José del Campo-Ávila,et al.  Online and Non-Parametric Drift Detection Methods Based on Hoeffding’s Bounds , 2015, IEEE Transactions on Knowledge and Data Engineering.

[5]  A. Bifet,et al.  Early Drift Detection Method , 2005 .

[6]  Gábor Lugosi,et al.  Prediction, learning, and games , 2006 .

[7]  Mohamed Medhat Gaber,et al.  Learning from Data Streams: Processing Techniques in Sensor Networks , 2007 .

[8]  Michèle Basseville,et al.  Detection of abrupt changes: theory and application , 1993 .

[9]  Stuart J. Russell,et al.  Online bagging and boosting , 2005, 2005 IEEE International Conference on Systems, Man and Cybernetics.

[10]  Kuei-Hu Chang,et al.  A more general risk assessment methodology using a soft set-based ranking technique , 2014, Soft Comput..

[11]  João Gama,et al.  Learning with Drift Detection , 2004, SBIA.

[12]  José del Campo-Ávila,et al.  Aprendiendo con detección de cambio online , 2014, Computación y Sistemas.

[13]  Geoff Holmes,et al.  MOA: Massive Online Analysis , 2010, J. Mach. Learn. Res..

[14]  Wenjia Wang,et al.  On diversity and accuracy of homogeneous and heterogeneous ensembles , 2007, Int. J. Hybrid Intell. Syst..

[15]  Ricard Gavaldà,et al.  Learning from Time-Changing Data with Adaptive Windowing , 2007, SDM.

[16]  Wei Fan,et al.  Systematic data selection to mine concept-drifting data streams , 2004, KDD.

[17]  Li Wan,et al.  Heterogeneous Ensemble for Feature Drifts in Data Streams , 2012, PAKDD.

[18]  Albert Bifet,et al.  Adaptive Stream Mining: Pattern Learning and Mining from Evolving Data Streams , 2010, Frontiers in Artificial Intelligence and Applications.

[19]  Geoff Holmes,et al.  New ensemble methods for evolving data streams , 2009, KDD.

[20]  Subhash C. Bagui,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2005, Technometrics.

[21]  Grigorios Tsoumakas,et al.  An Ensemble of Classifiers for coping with Recurring Contexts in Data Streams , 2008, ECAI.

[22]  João Gama,et al.  Issues in evaluation of stream learning algorithms , 2009, KDD.

[23]  Stuart J. Russell,et al.  Experimental comparisons of online and batch versions of bagging and boosting , 2001, KDD '01.

[24]  Claude Sammut,et al.  Extracting Hidden Context , 1998, Machine Learning.