A Robust Classifier Ensemble for Improving the Performance of Classification

Usage of recognition systems has found many applications in almost all fields. Generally in design of multiple classifier systems, the more diverse the results of the classifiers, the more appropriate the aggregated result. While most of classification algorithms have obtained a good performance for specific problems they have not enough robustness for other problems. Combination of multiple classifiers can be considered as a general solution method for pattern recognition problems. It has been shown that combination of multiple classifiers can usually operate better than a single classifier system provided that its components are independent or their components have diverse outputs. It has been shown that the necessary diversity for the ensemble can be achieved by manipulation of dataset features, manipulation of data points in dataset, different sub-samplings of dataset, and usage of different classification algorithms. We also propose a new method of creating this diversity. We use Linear Discriminant Analysis to manipulate the data points in dataset. The ensemble created by proposed method may not always outperform any of its members, it always possesses the diversity needed for creation of an ensemble, and consequently it always outperforms the simple classifier systems.

[1]  Hua Yu,et al.  A direct LDA algorithm for high-dimensional data - with application to face recognition , 2001, Pattern Recognit..

[2]  Zoran Obradovic,et al.  Effective pruning of neural network classifier ensembles , 2001, IJCNN'01. International Joint Conference on Neural Networks. Proceedings (Cat. No.01CH37222).

[3]  Simon Haykin,et al.  Neural Networks: A Comprehensive Foundation , 1998 .

[4]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1997, EuroCOLT.

[5]  William F. Punch,et al.  Optimizing Classification Ensembles via a Genetic Algorithm for a Web-Based Educational System , 2004, SSPR/SPR.

[6]  Hamid Parvin,et al.  An Innovative Feature Selection Using Fuzzy Entropy , 2011, ISNN.

[7]  Bruce E. Rosen,et al.  Ensemble Learning Using Decorrelated Neural Networks , 1996, Connect. Sci..

[8]  Jianxin Wu,et al.  Genetic Algorithm based Selective Neural Network Ensemble , 2001, IJCAI.

[9]  Fu Qiang PSO-based approach for neural network ensembles , 2004 .

[10]  Lior Rokach,et al.  Taxonomy for characterizing ensemble methods in classification tasks: A review and annotated bibliography , 2009, Comput. Stat. Data Anal..

[11]  Hamid Parvin,et al.  Linkage Learning Based on Local Optima , 2011, ICCCI.

[12]  Hamid Parvin,et al.  Linkage learning based on differences in local optimums of building blocks with one optima , 2011 .

[13]  Hamid Parvin,et al.  Detection of Cancer Patients Using an Innovative Method for Learning at Imbalanced Datasets , 2011, RSKT.

[14]  Hamid Parvin,et al.  A New N-gram Feature Extraction-Selection Method for Malicious Code , 2011, ICANNGA.

[15]  Anders Krogh,et al.  Neural Network Ensembles, Cross Validation, and Active Learning , 1994, NIPS.

[16]  Hamid Parvin,et al.  Localizing Program Logical Errors Using Extraction of Knowledge from Invariants , 2011, SEA.

[17]  傅强,et al.  Clustering-based selective neural network ensemble , 2005 .

[18]  Hamid Parvin,et al.  A Metric to Evaluate a Cluster by Eliminating Effect of Complement Cluster , 2011, KI.

[19]  David W. Opitz,et al.  Actively Searching for an E(cid:11)ective Neural-Network Ensemble , 1996 .

[20]  Chun-Xia Zhang,et al.  A local boosting algorithm for solving classification problems , 2008, Comput. Stat. Data Anal..

[21]  Hamid Parvin,et al.  A New Clustering Algorithm with the Convergence Proof , 2011, KES.

[22]  L. Breiman Arcing classifier (with discussion and a rejoinder by the author) , 1998 .

[23]  William F. Punch,et al.  Effects of resampling method and adaptation on clustering ensemble efficacy , 2011, Artificial Intelligence Review.

[24]  Hamid Parvin,et al.  A Novel Classifier Ensemble Method Based on Class Weightening in Huge Dataset , 2011, ISNN.

[25]  Shigeo Abe DrEng Pattern Classification , 2001, Springer London.

[26]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[27]  Yoav Freund,et al.  Experiments with a New Boosting Algorithm , 1996, ICML.

[28]  Louisa Lam,et al.  Classifier Combinations: Implementations and Theoretical Issues , 2000, Multiple Classifier Systems.

[29]  Xin Yao,et al.  Evolutionary ensembles with negative correlation learning , 2000, IEEE Trans. Evol. Comput..

[30]  Behrouz Minaei-Bidgoli,et al.  Multi objective association rule mining with genetic algorithm without specifying minimum support and minimum confidence , 2011, Expert Syst. Appl..

[31]  Lars Kai Hansen,et al.  Neural Network Ensembles , 1990, IEEE Trans. Pattern Anal. Mach. Intell..

[32]  Raymond J. Mooney,et al.  Constructing Diverse Classifier Ensembles using Artificial Training Examples , 2003, IJCAI.

[33]  David G. Stork,et al.  Pattern Classification , 1973 .

[34]  Horst Bunke,et al.  Creation of classifier ensembles for handwritten word recognition using feature selection algorithms , 2002, Proceedings Eighth International Workshop on Frontiers in Handwriting Recognition.

[35]  Hamid Parvin,et al.  On Possibility of Conditional Invariant Detection , 2011, KES.

[36]  Yuan-chin Ivan Chang,et al.  A stochastic approximation view of boosting , 2007, Comput. Stat. Data Anal..

[37]  Pablo M. Granitto,et al.  Selecting diverse members of neural network ensembles , 2000, Proceedings. Vol.1. Sixth Brazilian Symposium on Neural Networks.

[38]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[39]  L. Breiman Arcing Classifiers , 1998 .

[40]  Lior Rokach,et al.  Ensemble-based classifiers , 2010, Artificial Intelligence Review.

[41]  Subhash C. Bagui,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2005, Technometrics.

[42]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[43]  Hamid Parvin,et al.  An innovative combination of particle swarm optimization, learning automaton and great deluge algorithms for dynamic environments , 2011 .

[44]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.