SHAP and LIME: An Evaluation of Discriminative Power in Credit Risk

In credit risk estimation, the most important element is obtaining a probability of default as close as possible to the effective risk. This effort quickly prompted new, powerful algorithms that reach a far higher accuracy, but at the cost of losing intelligibility, such as Gradient Boosting or ensemble methods. These models are usually referred to as “black-boxes”, implying that you know the inputs and the output, but there is little way to understand what is going on under the hood. As a response to that, we have seen several different Explainable AI models flourish in recent years, with the aim of letting the user see why the black-box gave a certain output. In this context, we evaluate two very popular eXplainable AI (XAI) models in their ability to discriminate observations into groups, through the application of both unsupervised and predictive modeling to the weights these XAI models assign to features locally. The evaluation is carried out on real Small and Medium Enterprises data, obtained from official italian repositories, and may form the basis for the employment of such XAI models for post-processing features extraction.

[1]  J. MacQueen Some methods for classification and analysis of multivariate observations , 1967 .

[2]  Scott Lundberg,et al.  A Unified Approach to Interpreting Model Predictions , 2017, NIPS.

[3]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[4]  Paolo Giudici,et al.  Explainable AI in Fintech Risk Management , 2020, Frontiers in Artificial Intelligence.

[5]  Tianqi Chen,et al.  XGBoost: A Scalable Tree Boosting System , 2016, KDD.

[6]  Andreas Joseph,et al.  Parametric inference with universal function approximators , 2019, SSRN Electronic Journal.

[7]  Scott M. Lundberg,et al.  Consistent Individualized Feature Attribution for Tree Ensembles , 2018, ArXiv.

[8]  Andrew P. Bradley,et al.  The use of the area under the ROC curve in the evaluation of machine learning algorithms , 1997, Pattern Recognit..

[9]  P. Rousseeuw Silhouettes: a graphical aid to the interpretation and validation of cluster analysis , 1987 .

[10]  Paolo Giudici,et al.  Network based credit risk models , 2020, Quality Engineering.

[11]  Chih-Fong Tsai,et al.  Clustering-based undersampling in class-imbalanced data , 2017, Inf. Sci..

[12]  Joerg Osterrieder,et al.  Explainable AI in Credit Risk Management , 2021, SSRN Electronic Journal.

[13]  Javier Arroyo,et al.  Explainability of a Machine Learning Granting Scoring Model in Peer-to-Peer Lending , 2020, IEEE Access.

[14]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[15]  Michael I. Jordan,et al.  On Spectral Clustering: Analysis and an algorithm , 2001, NIPS.

[16]  Frank Zagare N-Person Games , 1984 .

[18]  Paolo Giudici,et al.  Lorenz Model Selection , 2020, J. Classif..

[19]  Paolo Giudici,et al.  Why to Buy Insurance? An Explainable Artificial Intelligence Approach , 2020, Risks.

[20]  Donald W. Bouldin,et al.  A Cluster Separation Measure , 1979, IEEE Transactions on Pattern Analysis and Machine Intelligence.