Performance Evaluation of the Machine Learning Algorithms Used in Inference Mechanism of a Medical Decision Support System

The importance of the decision support systems is increasingly supporting the decision making process in cases of uncertainty and the lack of information and they are widely used in various fields like engineering, finance, medicine, and so forth, Medical decision support systems help the healthcare personnel to select optimal method during the treatment of the patients. Decision support systems are intelligent software systems that support decision makers on their decisions. The design of decision support systems consists of four main subjects called inference mechanism, knowledge-base, explanation module, and active memory. Inference mechanism constitutes the basis of decision support systems. There are various methods that can be used in these mechanisms approaches. Some of these methods are decision trees, artificial neural networks, statistical methods, rule-based methods, and so forth. In decision support systems, those methods can be used separately or a hybrid system, and also combination of those methods. In this study, synthetic data with 10, 100, 1000, and 2000 records have been produced to reflect the probabilities on the ALARM network. The accuracy of 11 machine learning methods for the inference mechanism of medical decision support system is compared on various data sets.

[1]  Jie Cheng,et al.  Learning Bayesian Networks from Data: An Efficient Approach Based on Information Theory , 1999 .

[2]  Harold Borko,et al.  Encyclopedia of library and information science , 1970 .

[3]  David G. Stork,et al.  Pattern Classification , 1973 .

[4]  Wei-Yin Loh,et al.  Classification and regression trees , 2011, WIREs Data Mining Knowl. Discov..

[5]  John C. Platt,et al.  Fast training of support vector machines using sequential minimal optimization, advances in kernel methods , 1999 .

[6]  Tapio Elomaa,et al.  Selective Rademacher Penalization and Reduced Error Pruning of Decision Trees , 2004, J. Mach. Learn. Res..

[7]  Shigeo Abe DrEng Pattern Classification , 2001, Springer London.

[8]  S. G. Axline,et al.  An artificial intelligence program to advise physicians regarding antimicrobial therapy. , 1973, Computers and biomedical research, an international journal.

[9]  Edward H. Shortliffe,et al.  A model of inexact reasoning in medicine , 1990 .

[10]  Frederick Hayes-Roth,et al.  Rule-based systems , 1985, CACM.

[11]  Yingtao Jiang,et al.  A multilayer perceptron-based medical decision support system for heart disease diagnosis , 2006, Expert Syst. Appl..

[12]  R. Schapire The Strength of Weak Learnability , 1990, Machine Learning.

[13]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[14]  Geoffrey J. McLachlan,et al.  Ensemble Approach for the Classification of Imbalanced Data , 2009, Australasian Conference on Artificial Intelligence.

[15]  Alberto Maria Segre,et al.  Programs for Machine Learning , 1994 .

[16]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[17]  F. Cozman,et al.  Generalizing variable elimination in Bayesian networks , 2000 .

[18]  David J. Hand,et al.  An Empirical Comparison of Three Boosting Algorithms on Real Data Sets with Artificial Class Noise , 2003, Multiple Classifier Systems.

[19]  Carolyn M. Hall,et al.  Encyclopedia of Library and Information Science , 1971 .

[20]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[21]  Stephen M. Watt,et al.  Hybrid Mathematical Symbol Recognition Using Support Vector Machines , 2007, Ninth International Conference on Document Analysis and Recognition (ICDAR 2007).

[22]  Peter Reutemann,et al.  WEKA Manual for Version 3-6-10 , 2008 .

[23]  David J. Spiegelhalter,et al.  Local computations with probabilities on graphical structures and their application to expert systems , 1990 .

[24]  Tapio Elomaa,et al.  An Analysis of Reduced Error Pruning , 2001, J. Artif. Intell. Res..

[25]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[26]  Tshilidzi Marwala,et al.  Computational intelligence and decision trees for missing data estimation , 2008, 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence).

[27]  Shaogang Gong,et al.  Learning Support Vector Machines for A Multi-View Face Model , 1999, BMVC.

[28]  Enrique Frías-Martínez,et al.  Support vector machines versus multi-layer perceptrons for efficient off-line signature recognition , 2006, Eng. Appl. Artif. Intell..

[29]  John Langford,et al.  Machine Learning Techniques—Reductions Between Prediction Quality Metrics , 2008 .

[30]  Pat Langley,et al.  An Analysis of Bayesian Classifiers , 1992, AAAI.

[31]  Mark A. Hall,et al.  Correlation-based Feature Selection for Machine Learning , 2003 .

[32]  Somnuk Phon-Amnuaisuk,et al.  Using Rotation Forest for Protein Fold Prediction Problem: An Empirical Study , 2010, EvoBIO.

[33]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[34]  B. Webber,et al.  Artificial intelligence: a computerized decision aid for trauma. , 1988, The Journal of trauma.

[35]  Su-Yun Huang,et al.  Model selection for support vector machines via uniform design , 2007, Comput. Stat. Data Anal..

[36]  J. Friedman Special Invited Paper-Additive logistic regression: A statistical view of boosting , 2000 .

[37]  A Wakulicz-Deja,et al.  Diagnose progressive encephalopathy applying the rough set theory. , 1997, International journal of medical informatics.

[38]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[39]  Nobuhiko Yamaguchi Combining Pairwise Coupling Classifiers Using Individual Logistic Regressions , 2006, ICONIP.

[40]  J. Ross Quinlan,et al.  Simplifying decision trees , 1987, Int. J. Hum. Comput. Stud..

[41]  Harry Zhang,et al.  The Optimality of Naive Bayes , 2004, FLAIRS.

[42]  D. Kibler,et al.  Instance-based learning algorithms , 2004, Machine Learning.

[43]  A. H. Lav,et al.  Role of Learning Algorithm in Neural Network-Based Backcalculation of Flexible Pavements , 2006 .

[44]  Eibe Frank,et al.  Logistic Model Trees , 2003, Machine Learning.

[45]  Dominic Mazzoni,et al.  Multiclass reduced-set support vector machines , 2006, ICML.

[46]  Donato Malerba,et al.  A Comparative Analysis of Methods for Pruning Decision Trees , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[47]  P. Kokol,et al.  Decision trees and automatic learning in medical decision making , 1997, Proceedings Intelligent Information Systems. IIS'97.

[48]  Simon Kasif,et al.  OC1: A Randomized Induction of Oblique Decision Trees , 1993, AAAI.

[49]  Laura Schweitzer,et al.  Advances In Kernel Methods Support Vector Learning , 2016 .

[50]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[51]  Jan Martinovič,et al.  Creating of Conceptual Lattices using Multilayer Perceptron , 2005, CLA.

[52]  Hayri Sever,et al.  Comparison of different inference algorithms for medical decision making , 2014, Int. J. Comput. Intell. Syst..

[53]  Vladimir Nikulin Classification of Imbalanced Data with Random sets and Mean-Variance Filtering , 2008, Int. J. Data Warehous. Min..

[54]  David W. Aha,et al.  Noise-Tolerant Instance-Based Learning Algorithms , 1989, IJCAI.

[55]  Gregory F. Cooper,et al.  The ALARM Monitoring System: A Case Study with two Probabilistic Inference Techniques for Belief Networks , 1989, AIME.

[56]  Eibe Frank,et al.  Speeding Up Logistic Model Tree Induction , 2005, PKDD.