Explaining a machine learning decision to physicians via counterfactuals

Machine learning models perform well on several healthcare tasks and can help reduce the burden on the healthcare system. However, the lack of explainability is a major roadblock to their adoption in hospitals. \textit{How can the decision of an ML model be explained to a physician?} The explanations considered in this paper are counterfactuals (CFs), hypothetical scenarios that would have resulted in the opposite outcome. Specifically, time-series CFs are investigated, inspired by the way physicians converse and reason out decisions `I would have given the patient a vasopressor if their blood pressure was lower and falling'. Key properties of CFs that are particularly meaningful in clinical settings are outlined: physiological plausibility, relevance to the task and sparse perturbations. Past work on CF generation does not satisfy these properties, specifically plausibility in that realistic time-series CFs are not generated. A variational autoencoder (VAE)-based approach is proposed that captures these desired properties. The method produces CFs that improve on prior approaches quantitatively (more plausible CFs as evaluated by their likelihood w.r.t original data distribution, and 100$\times$ faster at generating CFs) and qualitatively (2$\times$ more plausible and relevant) as evaluated by three physicians.

[1]  A. Bouridane,et al.  Survey of Explainable AI Techniques in Healthcare , 2023, Sensors.

[2]  Chao Zhang,et al.  Counterfactual and Factual Reasoning over Hypergraphs for Interpretable Clinical Predictions on EHR , 2022, ML4H@NeurIPS.

[3]  D. Lasserson,et al.  The provision of hospital at home care: Results of a national survey of UK hospitals , 2021, International journal of clinical practice.

[4]  J. Schnipper,et al.  Hospital-Level Care at Home for Acutely Ill Adults: a Qualitative Evaluation of a Randomized Controlled Trial , 2021, Journal of General Internal Medicine.

[5]  Alessandro Blasimme,et al.  Explainability for artificial intelligence in healthcare: a multidisciplinary perspective , 2020, BMC Medical Informatics and Decision Making.

[6]  Anna Monreale,et al.  Explaining Any Time Series Classifier , 2020, 2020 IEEE Second International Conference on Cognitive Machine Intelligence (CogMI).

[7]  Mark T. Keane,et al.  Instance-Based Counterfactual Explanations for Time Series Classification , 2020, ICCBR.

[8]  Vitus J. Leung,et al.  Counterfactual Explanations for Multivariate Time Series , 2020, 2021 International Conference on Applied Artificial Intelligence (ICAPAI).

[9]  Oluwasanmi Koyejo,et al.  Towards Realistic Individual Recourse and Actionable Explanations in Black-Box Decision Making Systems , 2019, ArXiv.

[10]  Marzyeh Ghassemi,et al.  MIMIC-Extract: a data extraction, preprocessing, and representation pipeline for MIMIC-III , 2019, CHIL.

[11]  Cuntai Guan,et al.  A Survey on Explainable Artificial Intelligence (XAI): Toward Medical XAI , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[12]  Amit Sharma,et al.  Explaining machine learning classifiers through diverse counterfactual explanations , 2019, FAT*.

[13]  Kuangyan Song,et al.  "Why Should You Trust My Explanation?" Understanding Uncertainty in LIME Explanations , 2019 .

[14]  Klaus Kabino,et al.  Recursive neural networks in hospital bed occupancy forecasting , 2019, BMC Medical Informatics and Decision Making.

[15]  Jun S. Kim,et al.  An attention based deep learning model of clinical events in the intensive care unit , 2019, PloS one.

[16]  Masoumeh Haghpanahi,et al.  Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network , 2019, Nature Medicine.

[17]  Amina Adadi,et al.  Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.

[18]  Mohan S. Kankanhalli,et al.  Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda , 2018, CHI.

[19]  Amit Dhurandhar,et al.  Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives , 2018, NeurIPS.

[20]  A. Ng,et al.  CheXNet: Radiologist-Level Pneumonia Detection on Chest X-Rays with Deep Learning , 2017, ArXiv.

[21]  Roger G. Mark,et al.  Reproducibility in critical care: a mortality prediction case study , 2017, MLHC.

[22]  Peter Szolovits,et al.  Clinical Intervention Prediction and Understanding with Deep Neural Networks , 2017, MLHC.

[23]  Srikrishna Varadarajan,et al.  RADnet: Radiologist level accuracy using deep learning for hemorrhage detection in CT scans , 2017, 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018).

[24]  Chris Russell,et al.  Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.

[25]  Peter Szolovits,et al.  Predicting intervention onset in the ICU with switching state space models , 2017, CRI.

[26]  Jerrold H. May,et al.  Insights from a machine learning model for predicting the hospital Length of Stay (LOS) at the time of admission , 2017, Expert Syst. Appl..

[27]  Katherine A. Heller,et al.  Learning to Detect Sepsis with a Multitask Gaussian Process RNN Classifier , 2017, ICML.

[28]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[29]  Scott Lundberg,et al.  A Unified Approach to Interpreting Model Predictions , 2017, NIPS.

[30]  Been Kim,et al.  Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.

[31]  Carl Doersch,et al.  Tutorial on Variational Autoencoders , 2016, ArXiv.

[32]  Peter Szolovits,et al.  MIMIC-III, a freely accessible critical care database , 2016, Scientific Data.

[33]  P. Pronovost,et al.  A targeted real-time early warning score (TREWScore) for septic shock , 2015, Science Translational Medicine.

[34]  Max Welling,et al.  Auto-Encoding Variational Bayes , 2013, ICLR.

[35]  L. Deng,et al.  The MNIST Database of Handwritten Digit Images for Machine Learning Research [Best of the Web] , 2012, IEEE Signal Processing Magazine.

[36]  Donald Steinwachs,et al.  Hospital at Home: Feasibility and Outcomes of a Program To Provide Hospital-Level Care at Home for Acutely Ill Older Patients , 2005, Annals of Internal Medicine.

[37]  Gregory N. Hullender,et al.  Learning to rank using gradient descent , 2005, ICML.

[38]  Jeffrey M. Hausdorff,et al.  Physionet: Components of a New Research Resource for Complex Physiologic Signals". Circu-lation Vol , 2000 .