An Ensemble Method for Incremental Classification in Stationary and Non-stationary Environments

We present a model based on ensemble of base classifiers, that are combined using weighted majority voting, for the task of incremental classification. Definition of such voting weights becomes even more critical in non-stationary environments where the patterns underlying the observations change over time. Given an instance to classify, we propose to define each voting weight as a function that will take into account the location of an instance to classify in the different class-specific feature spaces and also the prior probability of such classes given the knowledge represented by the classifier as well as its overall performance in learning its training examples. This approach can improve the generalization performance and ability to control the stability/plasticity tradeoff, in stationary and non-stationary environments. Experiments were carried out using several real classification problems already introduced to test incremental algorithms in stationary as well as non-stationary environments.

[1]  James C. Bezdek,et al.  Decision templates for multiple classifier fusion: an experimental comparison , 2001, Pattern Recognit..

[2]  Fabio Roli,et al.  A theoretical and experimental analysis of linear combiners for multiple classifier systems , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[3]  R. Polikar,et al.  Ensemble based systems in decision making , 2006, IEEE Circuits and Systems Magazine.

[4]  R. Polikar,et al.  Dynamically weighted majority voting for incremental learning and comparison of three boosting based approaches , 2005, Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005..

[5]  Gerhard Widmer,et al.  Learning in the Presence of Concept Drift and Hidden Contexts , 1996, Machine Learning.

[6]  Saso Dzeroski,et al.  Combining Classifiers with Meta Decision Trees , 2003, Machine Learning.

[7]  Manfred K. Warmuth,et al.  The Weighted Majority Algorithm , 1994, Inf. Comput..

[8]  Jean-François Boulicaut,et al.  Local Pattern Detection , 2008 .

[9]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[10]  Stuart J. Russell,et al.  Online bagging and boosting , 2005, 2005 IEEE International Conference on Systems, Man and Cybernetics.

[11]  Subhash C. Bagui,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2005, Technometrics.

[12]  Martin Scholz,et al.  Knowledge-Based Sampling for Subgroup Discovery , 2004, Local Pattern Detection.

[13]  Robi Polikar,et al.  Learn++.MT: A New Approach to Incremental Learning , 2004, Multiple Classifier Systems.

[14]  Ralf Klinkenberg,et al.  Boosting classifiers for drifting concepts , 2007, Intell. Data Anal..

[15]  Alberto Prieto,et al.  Computational and ambient intelligence , 2009, Neurocomputing.

[16]  Claudio Moraga,et al.  Probabilistic Aggregation of Classifiers for Incremental Learning , 2007, IWANN.

[17]  Robert Givan,et al.  Online Ensemble Learning: An Empirical Study , 2000, Machine Learning.

[18]  Vasant Honavar,et al.  Learn++: an incremental learning algorithm for supervised neural networks , 2001, IEEE Trans. Syst. Man Cybern. Part C.

[19]  Ralf Klinkenberg,et al.  Learning drifting concepts: Example selection vs. example weighting , 2004, Intell. Data Anal..

[20]  Ludmila I. Kuncheva,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2004 .

[21]  Yoav Freund,et al.  A Short Introduction to Boosting , 1999 .