Diversity-Based Ensemble with Sample Weight Learning

Given multiple classifiers, one prevalent approach in classifier ensemble is to diversely combine classifier components (diversity-based ensemble), and a lot of previous works show that this approach can improve accuracy in classification. However, how to measure diversity and perform diversity-based learning are still challenges in the literature. Moreover, the learning procedure highly depends upon the distribution of the training data. In this paper, we propose a novel classifier ensemble method which combines classifiers with both diversity and sample weighting. First, by designing a matrix for the (sample) data distribution creatively, we formulate a unified optimization model for diversity-based ensemble with sample weighting, where classifier weights are learned through a convex quadratic programming problem with given sample weights. Second, we propose a new self-training algorithm to iteratively run the convex optimization and automatically learn the sample weights. Moreover, these sample weights are updated with a dynamically damped learning trick, which has a good performance for convergence. This paper also discusses the relationship between our optimization model and the margin theory. Extensive experiments on a variety of 50 UCI classification benchmark data sets show that the proposed approach consistently outperforms conventional ensembles such as Bagging, GASEN, and SDP.

[1]  Daniel Hernández-Lobato,et al.  An Analysis of Ensemble Pruning Techniques Based on Ordered Aggregation , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[2]  Zhi-Hua Zhou,et al.  On the doubt about margin explanation of boosting , 2010, Artif. Intell..

[3]  Yoav Freund,et al.  Experiments with a New Boosting Algorithm , 1996, ICML.

[4]  Kaizhu Huang,et al.  Convex ensemble learning with sparsity and diversity , 2014, Inf. Fusion.

[5]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[6]  William Nick Street,et al.  Ensemble Pruning Via Semi-definite Programming , 2006, J. Mach. Learn. Res..

[7]  Carlos Soares,et al.  A Comparison of Ranking Methods for Classification Algorithm Selection , 2000, ECML.

[8]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[9]  Kaizhu Huang,et al.  A novel classifier ensemble method with sparsity and diversity , 2014, Neurocomputing.

[10]  Zhi-Hua Zhou,et al.  Ensemble Methods: Foundations and Algorithms , 2012 .

[11]  Oscar Cordón,et al.  On the Combination of Accuracy and Diversity Measures for Genetic Selection of Bagging Fuzzy Rule-Based Multiclassification Systems , 2009, 2009 Ninth International Conference on Intelligent Systems Design and Applications.

[12]  Wei Tang,et al.  Ensembling neural networks: Many could be better than all , 2002, Artif. Intell..

[13]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[14]  H. Zou,et al.  NEW MULTICATEGORY BOOSTING ALGORITHMS BASED ON MULTICATEGORY FISHER-CONSISTENT LOSSES. , 2008, The annals of applied statistics.

[15]  Ludmila I. Kuncheva,et al.  Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy , 2003, Machine Learning.