Extended multimodal Eigenclassifiers and criteria for fusion model selection

Diversity among base classifiers is one of the key issues in classifier combination. Although the Eigenclassifiers method proposed by Ulas et al. (2012) aims to create uncorrelated base classifier outputs, however for multiclass classification problems, correlation among base classifier outputs arise due to the redundant features in the transformed classifier output space, which causes higher estimator variance and lower prediction accuracy. In this paper, we extend Eigenclassifiers method to obtain truly uncorrelated base classifiers. We also generalize the distribution on base classifier outputs from unimodal to multimodal, which lets us handle the class imbalance problem. We also aim to answer the question of which classifier fusion method should be used for a given dataset. In order to answer this question, we generate a dataset by calculating the performances of ten different fusion methods on 38 different datasets. We investigate accuracy-diversity relationship of ensembles on this experimental dataset by using eigenvalue distributions and divergence metrics defined by Kuncheva and Whitaker (2001). We obtain basic rules which can be used to decide on a fusion method given a dataset.

[1]  Hyun-Chul Kim,et al.  Bayesian Classifier Combination , 2012, AISTATS.

[2]  Ethem Alpaydin,et al.  Incremental construction of classifier and discriminant ensembles , 2009, Inf. Sci..

[3]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[4]  Ethem Alpaydin,et al.  Eigenclassifiers for combining correlated classifiers , 2012, Inf. Sci..

[5]  C. Eckart,et al.  The approximation of one matrix by another of lower rank , 1936 .

[6]  Yoav Freund,et al.  Experiments with a New Boosting Algorithm , 1996, ICML.

[7]  Nitish Srivastava,et al.  Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.

[8]  Louisa Lam,et al.  Classifier Combinations: Implementations and Theoretical Issues , 2000, Multiple Classifier Systems.

[9]  Mehmet Gönen,et al.  Bayesian Efficient Multiple Kernel Learning , 2012, ICML.

[10]  Bernhard Schölkopf,et al.  Kernel Principal Component Analysis , 1997, ICANN.

[11]  Paul Mineiro,et al.  Discriminative Features via Generalized Eigenvectors , 2013, ICML.

[12]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[13]  Fabio Roli,et al.  A theoretical and experimental analysis of linear combiners for multiple classifier systems , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[14]  C. J. Whitaker,et al.  Ten measures of diversity in classifier ensembles: limits for two classifiers , 2001 .

[15]  Thomas Hofmann,et al.  Data Integration for Classification Problems Employing Gaussian Process Priors , 2007 .

[16]  Ludmila I. Kuncheva,et al.  A Theoretical Study on Six Classifier Fusion Strategies , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[17]  Zehra Cataltepe,et al.  Classifier combination with kernelized eigenclassifiers , 2013, Proceedings of the 16th International Conference on Information Fusion.

[18]  Geoffrey E. Hinton,et al.  On the importance of initialization and momentum in deep learning , 2013, ICML.

[19]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[20]  Samy Bengio,et al.  How do correlation and variance of base-experts affect fusion in biometric authentication tasks? , 2005, IEEE Transactions on Signal Processing.

[21]  Mei-Chen Yeh,et al.  Multimodal fusion using learned text concepts for image categorization , 2006, MM '06.