Sensitivity Analysis of the Result in Binary Decision Trees

This paper proposes a new method to qualify the result given by a decision tree when it is used as a decision aid system. When the data are numerical, we compute the distance of a case from the decision surface. This distance measures the sensitivity of the result to a change in the input data. With a different distance it is also possible to measure the sensitivity of the result to small changes in the tree. The distance from the decision surface can also be combined to the error rate in order to provide a context-dependent information to the end-user.

[1]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[2]  Alberto Maria Segre,et al.  Programs for Machine Learning , 1994 .

[3]  Isabelle Alvarez,et al.  Explaining the Result of a Decision Tree to the End-User , 2004, ECAI.

[4]  Ron Kohavi,et al.  Scaling Up the Accuracy of Naive-Bayes Classifiers: A Decision-Tree Hybrid , 1996, KDD.

[5]  Heinz H. Bauschke,et al.  On Projection Algorithms for Solving Convex Feasibility Problems , 1996, SIAM Rev..

[6]  John Platt,et al.  Probabilistic Outputs for Support vector Machines and Comparisons to Regularized Likelihood Methods , 1999 .

[7]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques, 3rd Edition , 1999 .

[8]  Bojan Cestnik,et al.  Estimating Probabilities: A Crucial Task in Machine Learning , 1990, ECAI.

[9]  Donato Malerba,et al.  A Comparative Analysis of Methods for Pruning Decision Trees , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[10]  Tom Fawcett,et al.  Analysis and Visualization of Classifier Performance: Comparison under Imprecise Class and Cost Distributions , 1997, KDD.

[11]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[12]  Ron Kohavi,et al.  A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection , 1995, IJCAI.

[13]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques with Java implementations , 2002, SGMD.

[14]  Robert A. Jacobs,et al.  Hierarchical Mixtures of Experts and the EM Algorithm , 1993, Neural Computation.

[15]  Pedro M. Domingos A Unifeid Bias-Variance Decomposition and its Applications , 2000, ICML.

[16]  Alexander J. Smola,et al.  Advances in Large Margin Classifiers , 2000 .

[17]  Dana Ron,et al.  Algorithmic Stability and Sanity-Check Bounds for Leave-one-Out Cross-Validation , 1997, COLT.

[18]  David West,et al.  Neural network credit scoring models , 2000, Comput. Oper. Res..

[19]  Sreerama K. Murthy,et al.  Automatic Construction of Decision Trees from Data: A Multi-Disciplinary Survey , 1998, Data Mining and Knowledge Discovery.

[20]  Pedro M. Domingos A Unifeid Bias-Variance Decomposition and its Applications , 2000, ICML.

[21]  Pedro M. Domingos MetaCost: a general method for making classifiers cost-sensitive , 1999, KDD '99.

[22]  Bianca Zadrozny,et al.  Obtaining calibrated probability estimates from decision trees and naive Bayesian classifiers , 2001, ICML.

[23]  Dana Ron,et al.  Algorithmic Stability and Sanity-Check Bounds for Leave-One-Out Cross-Validation , 1997, Neural Computation.

[24]  D A Bloch,et al.  Recursive partitioning for the identification of disease risk subgroups: a case-control study of subarachnoid hemorrhage. , 1998, Journal of clinical epidemiology.

[25]  TreesKristin P. Bennett,et al.  Optimal Decision Trees , 1996 .

[26]  øöö Blockinøø Well-Trained PETs : Improving Probability Estimation , 2000 .