Nomograms for Visualization of Naive Bayesian Classifier

Besides good predictive performance, the naive Bayesian classifier can also offer a valuable insight into the structure of the training data and effects of the attributes on the class probabilities. This structure may be effectively revealed through visualization of the classifier. We propose a new way to visualize the naive Bayesian model in the form of a nomogram. The advantages of the proposed method are simplicity of presentation, clear display of the effects of individual attribute values, and visualization of confidence intervals. Nomograms are intuitive and when used for decision support can provide a visual explanation of predicted probabilities. And finally, with a nomogram, a naive Bayesian model can be printed out and used for probability prediction without the use of computer or calculator.

[1]  M. Kattan,et al.  A preoperative nomogram for disease recurrence following radical prostatectomy for prostate cancer. , 1998, Journal of the National Cancer Institute.

[2]  Ron Kohavi,et al.  Visualizing the simple Baysian classifier , 2001 .

[3]  B. Marx The Visual Display of Quantitative Information , 1985 .

[4]  Andreas Wierse,et al.  Information Visualization in Data Mining and Knowledge Discovery , 2001 .

[5]  David W. Hosmer,et al.  Applied Logistic Regression , 1991 .

[6]  J. Lubsen,et al.  A Practical Device for the Application of a Diagnostic or Prognostic Function , 1978, Methods of Information in Medicine.

[7]  Michael I. Jordan,et al.  On Discriminative vs. Generative Classifiers: A comparison of logistic regression and naive Bayes , 2001, NIPS.

[8]  Sunil J Rao,et al.  Regression Modeling Strategies: With Applications to Linear Models, Logistic Regression, and Survival Analysis , 2003 .

[9]  Igor Kononenko,et al.  Inductive and Bayesian learning in medical diagnosis , 1993, Appl. Artif. Intell..

[10]  Colin Ware,et al.  Information Visualization: Perception for Design , 2000 .

[11]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[12]  Pedro M. Domingos,et al.  Beyond Independence: Conditions for the Optimality of the Simple Bayesian Classifier , 1996, ICML.

[13]  Ivan Bratko,et al.  Analyzing Attribute Dependencies , 2003, PKDD.

[14]  Ivan Bratko,et al.  Machine Learning for Survival Analysis: A Case Study on Recurrence of Prostate Cancer , 1999, AIMDM.

[15]  Blaz Zupan,et al.  Orange: From Experimental Machine Learning to Interactive Data Mining , 2004, PKDD.

[16]  Bojan Cestnik,et al.  Estimating Probabilities: A Crucial Task in Machine Learning , 1990, ECAI.