Introducing randomness into greedy ensemble pruning algorithms

As is well known, the Greedy Ensemble Pruning (GEP) algorithm, also called the Directed Hill Climbing Ensemble Pruning (DHCEP) algorithm, possesses relatively good performance and high speed. However, because the algorithm only explores a relatively small subspace within the whole solution space, it often produces suboptimal solutions of the ensemble pruning problem. Aiming to address this drawback, in this work, we propose a novel Randomized GEP (RandomGEP) algorithm, also called the Randomized DHCEP (RandomDHCEP) algorithm, that effectively enlarges the search space of the classical DHCEP while maintaining the same level of time complexity with the help of a randomization technique. The randomization of the classical DHCEP algorithm achieves a good tradeoff between the effectiveness and efficiency of ensemble pruning. Besides, the RandomDHCEP algorithm naturally inherits the two intrinsic advantages that a randomized algorithm usually possesses. First, in most cases, its running time or space requirements are smaller than well-behaved deterministic ensemble pruning algorithms. Second, it is easy to comprehend and implement. Experimental results on three benchmark classification datasets verify the practicality and effectiveness of the RandomGEP algorithm.

[1]  Rich Caruana,et al.  Ensemble selection from libraries of models , 2004, ICML.

[2]  Dai Qun,et al.  Improved CBP Neural Network Model with Applications in Time Series Prediction , 2003 .

[3]  Thomas G. Dietterich An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization , 2000, Machine Learning.

[4]  Thomas G. Dietterich Multiple Classifier Systems , 2000, Lecture Notes in Computer Science.

[5]  A. J. M. M. Weijters The BP-SOM architecture and learning rule , 2006, Neural Processing Letters.

[6]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[7]  Salvatore J. Stolfo,et al.  Cost Complexity-Based Pruning of Ensemble Classifiers , 2001, Knowledge and Information Systems.

[8]  Qun Dai,et al.  The build of a dynamic classifier selection ICBP system and its application to pattern recognition , 2010, Neural Computing and Applications.

[9]  Wei Tang,et al.  Selective Ensemble of Decision Trees , 2003, RSFDGrC.

[10]  Qun Dai,et al.  The build of n-Bits Binary Coding ICBP Ensemble System , 2011, Neurocomputing.

[11]  H. Jaap van den Herik,et al.  Interpretable Neural Networks with BP-SOM , 1998, ECML.

[12]  Lawrence O. Hall,et al.  Ensemble diversity measures and their application to thinning , 2004, Inf. Fusion.

[13]  Leo Breiman,et al.  Pasting Small Votes for Classification in Large Databases and On-Line , 1999, Machine Learning.

[14]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[15]  Robert E. Schapire,et al.  The Boosting Approach to Machine Learning An Overview , 2003 .

[16]  Wei Tang,et al.  Ensembling neural networks: Many could be better than all , 2002, Artif. Intell..

[17]  Alberto Suárez,et al.  Aggregation Ordering in Bagging , 2004 .

[18]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[19]  Jeroen Eggermont,et al.  Rule-extraction and learning in the BP-SOM architecture , 1998 .

[20]  G DietterichThomas An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees , 2000 .

[21]  Grigorios Tsoumakas,et al.  An ensemble uncertainty aware measure for directed hill climbing ensemble pruning , 2010, Machine Learning.

[22]  Thomas G. Dietterich,et al.  Pruning Adaptive Boosting , 1997, ICML.

[23]  H. J. van den Herik,et al.  Intelligible neural networks with BP-SOM , 1997 .

[24]  Songcan Chen,et al.  Improved CBP Neural Network Model with Applications in Time Series Prediction , 2004, Neural Processing Letters.

[25]  Rich Caruana,et al.  Getting the Most Out of Ensemble Selection , 2006, Sixth International Conference on Data Mining (ICDM'06).

[26]  Yoav Freund,et al.  Experiments with a New Boosting Algorithm , 1996, ICML.

[27]  David H. Wolpert,et al.  Stacked generalization , 1992, Neural Networks.

[28]  Muhammad H. Alsuwaiyel,et al.  Algorithms - Design Techniques and Analysis , 1999, Lecture Notes Series on Computing.