Evidential multinomial logistic regression for multiclass classifier calibration

The calibration of classifiers is an important task in information fusion. To compare or combine the outputs of several classifiers, they need to be represented in a common space. Probabilistic calibration methods transform the output of a classifier into a posterior probability distribution. In this paper, we introduce an evidential calibration method for multiclass classification problems. Our approach uses an extension of multinomial logistic regression to the theory of belief functions. We demonstrate that the use of belief functions instead of probability distributions is often beneficial. In particular, when different classifiers are trained with unbalanced amount of training data, the gain achieved by our evidential approach can become significant. We applied our method to the calibration of multiclass SVM classifiers which were constructed through a “one-vs-all” framework. Experiments were conducted using six different datasets from the UCI repository.

[1]  Thierry Denoeux,et al.  Analysis of evidence-theoretic decision rules for pattern classification , 1997, Pattern Recognit..

[2]  Robert Sabourin,et al.  “One Against One” or “One Against All”: Which One is Better for Handwriting Recognition with SVMs? , 2006 .

[3]  Chih-Jen Lin,et al.  Probability Estimates for Multi-class Classification by Pairwise Coupling , 2003, J. Mach. Learn. Res..

[4]  Arthur P. Dempster,et al.  The Dempster-Shafer calculus for statisticians , 2008, Int. J. Approx. Reason..

[5]  Thierry Denoeux,et al.  Forecasting using belief functions: An application to marketing econometrics , 2014, Int. J. Approx. Reason..

[6]  Thierry Denux,et al.  Likelihood-based belief function: Justification and some extensions to low-quality data , 2014, Int. J. Approx. Reason..

[7]  Thierry Denoeux,et al.  Rejoinder on "Likelihood-based belief function: Justification and some extensions to low-quality data" , 2014, Int. J. Approx. Reason..

[8]  P. Walley Statistical Reasoning with Imprecise Probabilities , 1990 .

[9]  Sébastien Destercke,et al.  Correcting Binary Imprecise Classifiers: Local vs Global Approach , 2012, SUM.

[10]  Bianca Zadrozny,et al.  Obtaining calibrated probability estimates from decision trees and naive Bayesian classifiers , 2001, ICML.

[11]  Koby Crammer,et al.  On the Learnability and Design of Output Codes for Multiclass Problems , 2002, Machine Learning.

[12]  Glenn Shafer,et al.  A Mathematical Theory of Evidence , 2020, A Mathematical Theory of Evidence.

[13]  Benjamin Quost One-against-all Classifier Combination in the Framework of Belief Functions , 2006 .

[14]  Thierry Denoeux,et al.  Pairwise Classifier Combination in the Framework of Belief Functions , 2005 .

[15]  John Platt,et al.  Probabilistic Outputs for Support vector Machines and Comparisons to Regularized Likelihood Methods , 1999 .

[16]  Hung T. Nguyen,et al.  An Introduction to Random Sets , 2006 .

[17]  Thierry Denoeux,et al.  Evidential Logistic Regression for Binary SVM Classifier Calibration , 2014, Belief Functions.

[18]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[19]  Thierry Denœux Rejoinder on “Likelihood-based belief function: Justification and some extensions to low-quality data” , 2014 .

[20]  Robert Tibshirani,et al.  Classification by Pairwise Coupling , 1997, NIPS.

[21]  Bianca Zadrozny,et al.  Transforming classifier scores into accurate multiclass probability estimates , 2002, KDD.