Opening the black box of neural networks: methods for interpreting neural network models in clinical applications.

Artificial neural networks (ANNs) are powerful tools for data analysis and are particularly suitable for modeling relationships between variables for best prediction of an outcome. While these models can be used to answer many important research questions, their utility has been critically limited because the interpretation of the "black box" model is difficult. Clinical investigators usually employ ANN models to predict the clinical outcomes or to make a diagnosis; the model however is difficult to interpret for clinicians. To address this important shortcoming of neural network modeling methods, we describe several methods to help subject-matter audiences (e.g., clinicians, medical policy makers) understand neural network models. Garson's algorithm describes the relative magnitude of the importance of a descriptor (predictor) in its connection with outcome variables by dissecting the model weights. The Lek's profile method explores the relationship of the outcome variable and a predictor of interest, while holding other predictors at constant values (e.g., minimum, 20th quartile, maximum). While Lek's profile was developed specifically for neural networks, partial dependence plot is a more generic version that visualize the relationship between an outcome and one or two predictors. Finally, the local interpretable model-agnostic explanations (LIME) method can show the predictions of any classification or regression, by approximating it locally with an interpretable model. R code for the implementations of these methods is shown by using example data fitted with a standard, feed-forward neural network model. We offer codes and step-by-step description on how to use these tools to facilitate better understanding of ANN.

[1]  Kurt Hornik,et al.  Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.

[2]  G. David Garson,et al.  Interpreting neural-network connection weights , 1991 .

[3]  A. T. C. Goh,et al.  Back-propagation neural networks for modeling complex systems , 1995, Artif. Intell. Eng..

[4]  I. Dimopoulos,et al.  Application of neural networks to modelling nonlinear relationships in ecology , 1996 .

[5]  M. Gevrey,et al.  Review and comparison of methods to study the contribution of variables in artificial neural network models , 2003 .

[6]  Christina Gloeckner,et al.  Modern Applied Statistics With S , 2003 .

[7]  Bogusław Stefaniak,et al.  [Algorithms of artificial neural networks--practical application in medical science]. , 2005, Polski merkuriusz lekarski : organ Polskiego Towarzystwa Lekarskiego.

[8]  M. Defrise,et al.  Image reconstruction. , 2006, Physics in medicine and biology.

[9]  Hadley Wickham,et al.  ggplot2 - Elegant Graphics for Data Analysis (2nd Edition) , 2017 .

[10]  Holger R. Maier,et al.  Review of Input Variable Selection Methods for Artificial Neural Networks , 2011 .

[11]  Geoffrey J. Gordon,et al.  Artificial intelligence in medicine , 1989, Springer US.

[12]  Max Kuhn,et al.  caret: Classification and Regression Training , 2015 .

[13]  Sameer Singh,et al.  “Why Should I Trust You?”: Explaining the Predictions of Any Classifier , 2016, NAACL.

[14]  Jane C. Deng,et al.  Prediction of Clinical Deterioration in Hospitalized Adult Patients with Hematologic Malignancies Using a Neural Network Model , 2016, PloS one.

[15]  Eric Smalley,et al.  AI-powered drug discovery captures pharma interest , 2017, Nature Biotechnology.

[16]  Brandon M. Greenwell pdp: An R Package for Constructing Partial Dependence Plots , 2017, R J..

[17]  Marcus W Beck,et al.  NeuralNetTools: Visualization and Analysis Tools for Neural Networks. , 2018, Journal of statistical software.

[18]  Jae Yong Ryu,et al.  Deep learning improves prediction of drug–drug and drug–food interactions , 2018, Proceedings of the National Academy of Sciences.

[19]  Bruce R. Rosen,et al.  Image reconstruction by domain-transform manifold learning , 2017, Nature.