Comprehensible Credit Scoring Models Using Rule Extraction from Support Vector Machines

In recent years, Support Vector Machines (SVMs) were successfully applied to a wide range of applications. Their good performance is achieved by an implicit non-linear transformation of the original problem to a high-dimensional (possibly infinite) feature space in which a linear decision hyperplane is constructed that yields a nonlinear classifier in the input space. However, since the classifier is described as a complex mathematical function, it is rather incomprehensible for humans. This opacity property prevents them from being used in many real- life applications where both accuracy and comprehensibility are required, such as medical diagnosis and credit risk evaluation. To overcome this limitation, rules can be extracted from the trained SVM that are interpretable by humans and keep as much of the accuracy of the SVM as possible. In this paper, we will provide an overview of the recently proposed rule extraction techniques for SVMs and introduce two others taken from the artificial neural networks domain, being Trepan and G-REX. The described techniques are compared using publicly avail- able datasets, such as Ripley's synthetic dataset and the multi-class iris dataset. We will also look at medical diagnosis and credit scoring where comprehensibility is a key requirement and even a regulatory recommendation. Our experiments show that the SVM rule extraction techniques lose only a small percentage in performance compared to SVMs and therefore rank at the top of comprehensible classification techniques.

[1]  Joachim Diederich,et al.  Survey and critique of techniques for extracting rules from trained artificial neural networks , 1995, Knowl. Based Syst..

[2]  Johan A. K. Suykens,et al.  Faculteit Economie En Bedrijfskunde Hoveniersberg 24 B-9000 Gent Bayesian Kernel-based Classification for Financial Distress Detection Dirk Van Den Poel 4 Bayesian Kernel Based Classification for Financial Distress Detection , 2022 .

[3]  Harris Drucker,et al.  Support vector machines for spam categorization , 1999, IEEE Trans. Neural Networks.

[4]  South Africa MOODY'S KMV RISKCALC , 2005 .

[5]  Andreu Català,et al.  Rule Based Learning Systems from SVM and RBFNN , 2004 .

[6]  Lars Niklasson,et al.  The Truth is In There - Rule Extraction from Opaque Models Using Genetic Programming , 2004, FLAIRS.

[7]  Michael V. Mannino,et al.  The cost-minimizing inverse classification problem: a genetic algorithm approach , 2000, Decis. Support Syst..

[8]  Johan A. K. Suykens,et al.  Benchmarking state-of-the-art classification algorithms for credit scoring , 2003, J. Oper. Res. Soc..

[9]  P. J. Green,et al.  Density Estimation for Statistics and Data Analysis , 1987 .

[10]  Antony Browne,et al.  Biological data mining with neural networks: implementation and application of a flexible decision tree extraction algorithm to genomic problem domains , 2004, Neurocomputing.

[11]  Joachim Diederich,et al.  Learning-Based Rule-Extraction From Support Vector Machines: Performance On Benchmark Data Sets , 2004 .

[12]  Sabine Van Huffel,et al.  Preoperative prediction of malignancy of ovarian tumors using least squares support vector machines , 2003, Artif. Intell. Medicine.

[13]  Joachim Diederich,et al.  Rule Extraction from Support Vector Machines , 2008, Studies in Computational Intelligence.

[14]  Masato Koda,et al.  Sensitivity analysis in data mining , 1998 .

[15]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[16]  Mark Craven,et al.  Extracting comprehensible models from trained neural networks , 1996 .

[17]  Bart Baesens,et al.  Using Neural Network Rule Extraction and Decision Tables for Credit - Risk Evaluation , 2003, Manag. Sci..

[18]  Brian D. Ripley,et al.  Neural Networks and Related Methods for Classification , 1994 .

[19]  Glenn Fung,et al.  Rule extraction from linear support vector machines , 2005, KDD '05.

[20]  Alberto Maria Segre,et al.  Programs for Machine Learning , 1994 .

[21]  Nello Cristianini,et al.  An Introduction to Support Vector Machines and Other Kernel-based Learning Methods , 2000 .

[22]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[23]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.