Combined classifier based on feature space partitioning

This paper presents a significant modification to the AdaSS (Adaptive Splitting and Selection) algorithm, which was developed several years ago. The method is based on the simultaneous partitioning of the feature space and an assignment of a compound classifier to each of the subsets. The original version of the algorithm uses a classifier committee and a majority voting rule to arrive at a decision. The proposed modification replaces the fairly simple fusion method with a combined classifier, which makes a decision based on a weighted combination of the discriminant functions of the individual classifiers selected for the committee. The weights mentioned above are dependent not only on the classifier identifier, but also on the class number. The proposed approach is based on the results of previous works, where it was proven that such a combined classifier method could achieve significantly better results than simple voting systems. The proposed modification was evaluated through computer experiments, carried out on diverse benchmark datasets. The results are very promising in that they show that, for most of the datasets, the proposed method outperforms similar techniques based on the clustering and selection approach.

[1]  C. K. Chow,et al.  Statistical Independence and Threshold Functions , 1965, IEEE Trans. Electron. Comput..

[2]  D. Wolpert The Supervised Learning No-Free-Lunch Theorems , 2002 .

[3]  David E. Goldberg,et al.  Genetic Algorithms in Search Optimization and Machine Learning , 1988 .

[4]  Ludmila I. Kuncheva,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2004 .

[5]  Adam Krzyżak,et al.  Methods of combining multiple classifiers and their applications to handwriting recognition , 1992, IEEE Trans. Syst. Man Cybern..

[6]  Michal Wozniak Experiments on Linear Combiners , 2008, Information Technologies in Biomedicine.

[7]  Ofer MatanDepartment On Voting Ensembles of Classiiers (extended Abstract) , 1996 .

[8]  Michal Wozniak,et al.  Some Remarks on Chosen Methods of Classifier Fusion Based on Weighted Voting , 2009, HAIS.

[9]  Robert P. W. Duin,et al.  The combining classifier: to train or not to train? , 2002, Object recognition supported by user interaction for service robots.

[10]  Michal Wozniak,et al.  Optimization of overlay distributed computing systems for multiple classifier system - heuristic approach , 2012, Log. J. IGPL.

[11]  Mario Vento,et al.  A Cascaded Multiple Expert System for Verification , 2000, Multiple Classifier Systems.

[12]  Tin Kam Ho,et al.  The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  Geoffrey I. Webb,et al.  Feature-subspace aggregating: ensembles for stable and unstable learners , 2011, Machine Learning.

[14]  Robert Tibshirani,et al.  An Introduction to the Bootstrap , 1994 .

[15]  Ludmila I. Kuncheva,et al.  Clustering-and-selection model for classifier combination , 2000, KES'2000. Fourth International Conference on Knowledge-Based Intelligent Engineering Systems and Allied Technologies. Proceedings (Cat. No.00TH8516).

[16]  Andrzhei︠a︡ Mostowski,et al.  Set theory : with an introduction to descriptive set theory , 1976 .

[17]  Geoffrey E. Hinton,et al.  Adaptive Mixtures of Local Experts , 1991, Neural Computation.

[18]  Fabio Roli,et al.  Design of effective multiple classifier systems by clustering of classifiers , 2000, Proceedings 15th International Conference on Pattern Recognition. ICPR-2000.

[19]  Bartosz Krawczyk,et al.  Designing Cost-Sensitive Ensemble - Genetic Approach , 2011, IP&C.

[20]  Fabio Roli,et al.  Bayesian Analysis of Linear Combiners , 2007, MCS.

[21]  Bogdan Gabrys,et al.  Classifier selection for majority voting , 2005, Inf. Fusion.

[22]  Robert P. W. Duin,et al.  Limits on the majority vote accuracy in classifier fusion , 2003, Pattern Analysis & Applications.

[23]  R Core Team,et al.  R: A language and environment for statistical computing. , 2014 .

[24]  Derek Partridge,et al.  Software Diversity: Practical Statistics for Its Measurement and Exploitation | Draft Currently under Revision , 1996 .

[25]  Robert A. Jacobs,et al.  Methods For Combining Experts' Probability Assessments , 1995, Neural Computation.

[26]  Anil K. Jain,et al.  Data clustering: a review , 1999, CSUR.

[27]  Daniel A. Ashlock,et al.  Evolutionary computation for modeling and optimization , 2005 .

[28]  James C. Bezdek,et al.  Decision templates for multiple classifier fusion: an experimental comparison , 2001, Pattern Recognit..

[29]  R. Polikar,et al.  Ensemble based systems in decision making , 2006, IEEE Circuits and Systems Magazine.

[30]  K. Goebel Choosing Classifiers for Decision Fusion , 2004 .

[31]  Grzegorz Chmaj,et al.  Heuristic algorithms for optimization of task allocation and result distribution in peer-to-peer computing systems , 2012, Int. J. Appl. Math. Comput. Sci..

[32]  Emilio Corchado,et al.  Hybrid Classification Ensemble Using Topology-preserving Clustering , 2011, New Generation Computing.

[33]  Ethem Alpaydın,et al.  Combined 5 x 2 cv F Test for Comparing Supervised Classification Learning Algorithms , 1999, Neural Comput..

[34]  Anil K. Jain,et al.  Statistical Pattern Recognition: A Review , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[35]  Kevin W. Bowyer,et al.  Combination of Multiple Classifiers Using Local Accuracy Estimates , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[36]  Subhash C. Bagui,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2005, Technometrics.

[37]  Michal Wozniak,et al.  Algorithm of designing compound recognition system on the basis of combining classifiers with simultaneous splitting feature space into competence areas , 2009, Pattern Analysis and Applications.

[38]  J. MacQueen Some methods for classification and analysis of multivariate observations , 1967 .

[39]  Michal Wozniak,et al.  Complexity and Multithreaded Implementation Analysis of One Class-Classifiers Fuzzy Combiner , 2011, HAIS.

[40]  Bogdan Trawinski,et al.  Selection of Heterogeneous Fuzzy Model Ensembles Using Self-adaptive Genetic Algorithms , 2011, New Generation Computing.

[41]  Thomas G. Dietterich,et al.  Solving Multiclass Learning Problems via Error-Correcting Output Codes , 1994, J. Artif. Intell. Res..

[42]  Ethem Alpaydin,et al.  Introduction to machine learning , 2004, Adaptive computation and machine learning.

[43]  Yoram Baram,et al.  Partial Classification: The Benefit of Deferred Decision , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[44]  Lars Kai Hansen,et al.  Neural Network Ensembles , 1990, IEEE Trans. Pattern Anal. Mach. Intell..

[45]  Shigeo Abe DrEng Pattern Classification , 2001, Springer London.

[46]  Louis Vuurpijl,et al.  An overview and comparison of voting methods for pattern recognition , 2002, Proceedings Eighth International Workshop on Frontiers in Handwriting Recognition.

[47]  David G. Stork,et al.  Pattern Classification , 1973 .

[48]  Gian Luca Marcialis,et al.  Fusion of Face Recognition Algorithms for Video-Based Surveillance Systems , 2003 .

[49]  Olgierd Unold,et al.  Self-adaptation of parameters in a learning classifier system ensemble machine , 2010, Int. J. Appl. Math. Comput. Sci..

[50]  Krzysztof Walkowiak,et al.  Anycasting in connection-oriented computer networks: Models, algorithms and results , 2010, Int. J. Appl. Math. Comput. Sci..

[51]  Michał Woźniak,et al.  Combining classifiers using trained fuser - analytical and experimental results / , 2010 .

[52]  Lalit M. Patnaik,et al.  Genetic algorithms: a survey , 1994, Computer.

[53]  Gavin Brown,et al.  "Good" and "Bad" Diversity in Majority Vote Ensembles , 2010, MCS.

[54]  Kagan Tumer,et al.  Analysis of decision boundaries in linearly combined neural classifiers , 1996, Pattern Recognit..