A framework for probabilistic combination of multiple classifiers at an abstract level

Abstract In previous approaches, the combination of multiple classifiers depends heavily on one of the three classification results; measurement scores (measurement level), ranking (rank level), and top choice (abstract level). For a more general combination of multiple classifiers, it is desirable that combination methods should be developed at the abstract level. In combining multiple classifiers at this level, most studies have assumed that classifiers behave independently. Such an assumption degrades and biases the classification performance, in cases where highly dependent classifiers are added. In order to overcome such weaknesses, it should be possible to combine multiple classifiers in a probabilistic framework, using a Bayesian formalism. A probabilistic combination of multiple decisions of K classifiers needs a ( K + 1)st-order probability distribution. However, it is well known that such a distribution will become unmanageable to store and estimate, even for a small K . In this paper, a framework is proposed to optimally identify a product set of kth-order dependencies, where 1≤ k ⪯- K for the product approximation of the ( K + 1)st-order probability distribution from training samples, and to probabilistically combine multiple decisions by the identified product set, using the Bayesian formalism. This framework was tested and evaluated using a standardized CENPARMI data base. The results showed superior performance over other combination methods.

[1]  R. Gallager Information Theory and Reliable Communication , 1968 .

[2]  C. N. Liu,et al.  Approximating discrete probability distributions with dependence trees , 1968, IEEE Trans. Inf. Theory.

[3]  Philip M. Lewis,et al.  Approximating Probability Distributions to Reduce Storage Requirements , 1959, Information and Control.

[4]  Adam Krzyżak,et al.  Methods of combining multiple classifiers and their applications to handwriting recognition , 1992, IEEE Trans. Syst. Man Cybern..

[5]  E. Mandler,et al.  Combining the Classification Results of Independent Classifiers Based on the Dempster/Shafer Theory of Evidence , 1988 .

[6]  B. John Oommen,et al.  On using the chi-squared metric for determining stochastic dependence , 1992, Pattern Recognit..

[7]  Solomon Kullback,et al.  Approximating discrete probability distributions , 1969, IEEE Trans. Inf. Theory.

[8]  L. N. Kanal,et al.  Uncertainty in Artificial Intelligence 5 , 1990 .

[9]  J. Franke,et al.  A comparison of two approaches for combining the votes of cooperating classifiers , 1992, Proceedings., 11th IAPR International Conference on Pattern Recognition. Vol.II. Conference B: Pattern Recognition Methodology and Systems.

[10]  Ching Y. Suen,et al.  A Method of Combining Multiple Experts for the Recognition of Unconstrained Handwritten Numerals , 1995, IEEE Trans. Pattern Anal. Mach. Intell..

[11]  Jin Hyung Kim,et al.  Dependency Relationship Based Decision Combination in Multiple Classifier Systems , 1995, IJCAI.