System of Boosting Voting with Multiple Learning Algorithms

The Ensemble classifier has been an active research topic in the area of machine learning. In a classifica - tion task, the ensemble scheme determines a final class label from several individual results which are usually generated by several individual classifiers according to two major principles. The first is to use multiple learning algorithms to form the set of individual outcomes, and the second is to train a set of data fractions, which are generated from a given training dataset, with a weak learning algorithm to generate the set of results. The advantage of the first principle is on classification stabil- ity, whereas that of the second principle is on the accuracy improvement in terms of classification accuracy. Thus far, most studies in the literature have been based on either a study of the two principles of ensemble. In this paper we propose a new ensemble scheme to combine the two principles simul- taneously. We evaluate the performance on a classification task. Experimental results on several UCI benchmark datasets demonstrate that the proposed framework achieves improved performance in terms of classification accuracy compared to the conventional approaches.

[1]  Yoram Singer,et al.  Improved Boosting Algorithms Using Confidence-rated Predictions , 1998, COLT' 98.

[2]  Anand M. Narasimhamurthy Theoretical bounds of majority voting performance for a binary classification problem , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[3]  Huimin Zhao,et al.  Incorporating domain knowledge into data mining classifiers: An application in indirect lending , 2008, Decis. Support Syst..

[4]  Lawrence O. Hall,et al.  A Comparison of Decision Tree Ensemble Creation Techniques , 2007 .

[5]  Philip S. Yu,et al.  Mining concept-drifting data streams using ensemble classifiers , 2003, KDD '03.

[6]  Ludmila I. Kuncheva,et al.  Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy , 2003, Machine Learning.

[7]  Lawrence O. Hall,et al.  A Comparison of Decision Tree Ensemble Creation Techniques , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[8]  Juan José Rodríguez Diez,et al.  Rotation Forest: A New Classifier Ensemble Method , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[9]  Sung-Bae Cho,et al.  Neural Network Ensemble with Negatively Correlated Features for Cancer Classification , 2003, ICANN.

[10]  Charles Elkan,et al.  Boosting and Naive Bayesian learning , 1997 .

[11]  Ludmila I. Kuncheva,et al.  A Theoretical Study on Six Classifier Fusion Strategies , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[12]  Eric Bauer,et al.  An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants , 1999, Machine Learning.

[13]  Dariusz Plewczynski,et al.  Assessing Different Classification Methods for Virtual Screening , 2006, J. Chem. Inf. Model..

[14]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[15]  Robert P. W. Duin,et al.  Limits on the majority vote accuracy in classifier fusion , 2003, Pattern Analysis & Applications.

[16]  Gonzalo Martínez-Muñoz,et al.  Switching class labels to generate classification ensembles , 2005, Pattern Recognit..

[17]  Huan‐Xiang Zhou,et al.  Prediction of solvent accessibility and sites of deleterious mutations from protein sequence , 2005, Nucleic acids research.

[18]  Yoav Freund,et al.  Experiments with a New Boosting Algorithm , 1996, ICML.

[19]  Ching Y. Suen,et al.  Application of majority voting to pattern recognition: an analysis of its behavior and performance , 1997, IEEE Trans. Syst. Man Cybern. Part A.

[20]  Yang Wang,et al.  Boosting an associative classifier , 2006, IEEE Transactions on Knowledge and Data Engineering.

[21]  Etienne Barnard,et al.  Data characteristics that determine classifier performance , 2006 .

[22]  Yoshua Bengio,et al.  Boosting Neural Networks , 2000, Neural Computation.

[23]  Ana I. González Acuña An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, Boosting, and Randomization , 2012 .

[24]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[25]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[26]  Oleksandr Makeyev,et al.  Neural network with ensembles , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).