Ensemble feature selection with the simple Bayesian classification in medical diagnostics

Ensembles of simple Bayesian classifiers have traditionally not been in the focus of classification research partly because of the stability of a simple Bayesian classifier and because of the rarely valid basic assumption that the classification features are independent of each other, given the predicted value. As a way to try to circumvent these problems we suggest the use of an ensemble of simple Bayesian classifiers each concentrating on solving a sub-problem of the problem domain. Our experiments with the problem of separating acute appendicitis show that in this way it is possible to retain the comprehensibility and at the same time to increase the diagnostic accuracy, sensitivity, and specificity. The advantages of the approach include also simplicity and speed of learning, small storage space needed during the classification, speed of classification, and the possibility of incremental learning.

[1]  Ron Kohavi,et al.  A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection , 1995, IJCAI.

[2]  Charles Elkan,et al.  Boosting and Naive Bayesian learning , 1997 .

[3]  Thomas Richardson,et al.  Interpretable Boosted Naïve Bayes Classification , 1998, KDD.

[4]  Alexey Tsymbal,et al.  Ensemble Feature Selection with Dynamic Integration of Classifiers , 2001 .

[5]  Peter Kokol,et al.  Comparison of Three Databases with a Decision Tree Approach in the Medical Field of Acute Appendicitis , 2001, MedInfo.

[6]  Eric Bauer,et al.  An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants , 1999, Machine Learning.

[7]  Ted Pedersen,et al.  A Simple Approach to Building Ensembles of Naive Bayesian Classifiers for Word Sense Disambiguation , 2000, ANLP.

[8]  Pedro M. Domingos,et al.  On the Optimality of the Simple Bayesian Classifier under Zero-One Loss , 1997, Machine Learning.

[9]  Alexey Tsymbal,et al.  A Dynamic Integration Algorithm for an Ensemble of Classifiers , 1999, ISMIS.

[10]  Alexey Tsymbal,et al.  Feature Selection for Ensembles of Simple Bayesian Classifiers , 2002, ISMIS.

[11]  Tin Kam Ho,et al.  The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[12]  David W. Opitz,et al.  Feature Selection for Ensembles , 1999, AAAI/IAAI.

[13]  Ron Kohavi,et al.  Data mining using /spl Mscr//spl Lscr//spl Cscr/++ a machine learning library in C++ , 1996, Proceedings Eighth IEEE International Conference on Tools with Artificial Intelligence.

[14]  Lars Kai Hansen,et al.  Neural Network Ensembles , 1990, IEEE Trans. Pattern Anal. Mach. Intell..