Transduction and typicalness for quality assessment of individual classifications in machine learning and data mining

In the past, machine learning algorithms have been successfully used in many problems, and are emerging as valuable data analysis tools. However, their serious practical use is affected by the fact, that more often than not, they cannot produce reliable and unbiased assessments of their predictions' quality. In last years, several approaches for estimating reliability or confidence of individual classifiers have emerged, many of them building upon the algorithmic theory of randomness, such as (historically ordered) transduction-based confidence estimation, typicalness-based confidence estimation, and transductive reliability estimation. Unfortunately, they all have weaknesses: either they are tightly bound with particular learning algorithms, or the interpretation of reliability estimations is not always consistent with statistical confidence levels. In the paper, we propose a joint approach that compensates the mentioned weaknesses by integrating typicalness-based confidence estimation and transductive reliability estimation into joint confidence machine. The resulting confidence machine produces confidence values in the statistical sense (e.g., a confidence level of 95% means that in 95% the predicted class is also a true class), as well as provides us with a general principle that is independent of to the particular underlying classifier. We perform a series of tests with several different machine learning algorithms in several problem domains. We compare our results with that of a proprietary TCM-NN method as well as with kernel density estimation. We show that the proposed method significantly outperforms density estimation methods, and how it may be used to improve their performance.

[1]  Igor Kononenko,et al.  Semi-Naive Bayesian Classifier , 1991, EWSL.

[2]  Alexander Gammerman,et al.  Computationally Efficient Transductive Machines , 2000, ALT.

[3]  Matthew P. Wand,et al.  Kernel Smoothing , 1995 .

[4]  D. F. Specht,et al.  Experience with adaptive probabilistic neural networks and adaptive general regression neural networks , 1994, Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94).

[5]  Vladimir Vovk,et al.  Comparing the Bayes and Typicalness Frameworks , 2001, ECML.

[6]  Brian D. Ripley,et al.  Modern Applied Statistics with S Fourth edition , 2002 .

[7]  Hilan Bensusan,et al.  Meta-Learning by Landmarking Various Learning Algorithms , 2000, ICML.

[8]  Alison L Gibbs,et al.  On Choosing and Bounding Probability Metrics , 2002, math/0209021.

[9]  Philipp Slusallek,et al.  Introduction to real-time ray tracing , 2005, SIGGRAPH Courses.

[10]  Harry Wechsler,et al.  Transductive confidence machine for active learning , 2003, Proceedings of the International Joint Conference on Neural Networks, 2003..

[11]  Eric R. Ziegel,et al.  The Elements of Statistical Learning , 2003, Technometrics.

[12]  Stephen D. Bay,et al.  Characterizing Model Errors and Di erences , 2000 .

[13]  Johannes Fürnkranz,et al.  An Evaluation of Grading Classifiers , 2001, IDA.

[14]  Igor Kononenko,et al.  Reliable Classifications with Machine Learning , 2002, ECML.

[15]  Alexander Gammerman,et al.  Transductive Confidence Machines for Pattern Recognition , 2002, ECML.

[16]  Matjaz Kukar,et al.  Transductive reliability estimation for medical diagnosis , 2003, Artif. Intell. Medicine.

[17]  D. Rumelhart Parallel Distributed Processing Volume 1: Foundations , 1987 .

[18]  Alexander Gammerman,et al.  Transduction with Confidence and Credibility , 1999, IJCAI.

[19]  Calvin L. Williams,et al.  Modern Applied Statistics with S-Plus , 1997 .

[20]  J. R. Koehler,et al.  Modern Applied Statistics with S-Plus. , 1996 .

[21]  Alexander G. Gray,et al.  Retrofitting Decision Tree Classifiers Using Kernel Density Estimation , 1995, ICML.

[22]  Marko Robnik-Sikonja,et al.  Overcoming the Myopia of Inductive Learning Algorithms with RELIEFF , 2004, Applied Intelligence.

[23]  Pat Langley,et al.  Estimating Continuous Distributions in Bayesian Classifiers , 1995, UAI.

[24]  O. M. Halck,et al.  Using Hard Classifiers to Estimate Conditional Class Probabilities , 2002, ECML.