Evaluating the performance of classifiers is not as trivial as it would seem at a first glance. Even the most widely used methods such as measuring accuracy or error rate on a test set has severe limitations. Two of the most prominent limitations of these measures are that they do not consider misclassification costs and can be misleading when the classes have very different prior probabilities. On the last years, several researches have pointed out alternative methods to evaluate the performance of learning systems. Some of those methods are based on graphical evaluation of classifiers. Usually, a graphical evaluation lets the user analyze the performance of a classifier under different scenarios, for instance, with different misclassification costs, and to select the classifier parameters setting that provides the best result. The objective of this paper is to survey some of the most used graphical methods for performance evaluation, which do not rely on precise class and cost distribution information.
[1]
Charles Elkan,et al.
The Foundations of Cost-Sensitive Learning
,
2001,
IJCAI.
[2]
Thomas G. Dietterich,et al.
Improved Class Probability Estimates from Decision Tree Models
,
2003
.
[3]
Sholom M. Weiss,et al.
Computer Systems That Learn
,
1990
.
[4]
David J. Hand,et al.
Construction and Assessment of Classification Rules
,
1997
.
[5]
Pedro M. Domingos.
MetaCost: a general method for making classifiers cost-sensitive
,
1999,
KDD '99.
[6]
Tom Fawcett,et al.
Analysis and Visualization of Classifier Performance: Comparison under Imprecise Class and Cost Distributions
,
1997,
KDD.
[7]
Charles X. Ling,et al.
Data Mining for Direct Marketing: Problems and Solutions
,
1998,
KDD.
[8]
Leo Breiman,et al.
Classification and Regression Trees
,
1984
.
[9]
Pedro M. Domingos,et al.
Tree Induction for Probability-Based Ranking
,
2003,
Machine Learning.