Study of selective ensemble learning method and its diversity based on decision tree and neural network

Diversity among base classifiers is known to be a necessary condition for improving ensemble learning performance. In this paper, methods of selective ensemble learning including hill-climbing selection, ensemble forward sequential selection, ensemble backward sequential selection and clustering selection are studied. To measure the diversity among base classifiers in ensemble learning, the entropy E is selected as measuring method of diversity. The results of experiment show that classifiers which have the highest diversity are obtained using selective methods, and the ensemble performance is superior to the best single classifier. In addition, the classifiers selected by clustering selective technology also have the above characteristics, and the changes of the diversity are smaller when the accuracy has smaller fluctuations. Meanwhile, the number of clusters also impacts on the ensemble performance.

[1]  Li Kai,et al.  A Selective Approach to Neural Network Ensemble Based on Clustering Technology , 2005 .

[2]  Chen Nian-yi Clustering Algorithm Based Selective Ensemble , 2004 .

[3]  Hou-Kuan Huang,et al.  A selective approach to neural network ensemble based on clustering technology , 2004, Proceedings of 2004 International Conference on Machine Learning and Cybernetics (IEEE Cat. No.04EX826).

[4]  Zbigniew Suraj,et al.  Feature Selection Algorithm for Multiple Classifier Systems: A Hybrid Approach , 2008, Fundam. Informaticae.

[5]  Alexey Tsymbal,et al.  Ensemble feature selection with the simple Bayesian classification , 2003, Inf. Fusion.

[6]  Mykola Pechenizkiy,et al.  Diversity in search strategies for ensemble feature selection , 2005, Inf. Fusion.

[7]  Robert Sabourin,et al.  A dynamic overproduce-and-choose strategy for the selection of classifier ensembles , 2008, Pattern Recognit..

[8]  Grigorios Tsoumakas,et al.  Pruning an ensemble of classifiers via reinforcement learning , 2009, Neurocomputing.

[9]  Fabio Roli,et al.  Design of effective neural network ensembles for image classification purposes , 2001, Image Vis. Comput..

[10]  Ludmila I. Kuncheva,et al.  Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy , 2003, Machine Learning.

[11]  Robert Sabourin,et al.  Overfitting cautious selection of classifier ensembles with genetic algorithms , 2009, Inf. Fusion.

[12]  Gonzalo Martínez-Muñoz,et al.  Using boosting to prune bagging ensembles , 2007, Pattern Recognit. Lett..

[13]  Wei Tang,et al.  Ensembling neural networks: Many could be better than all , 2002, Artif. Intell..