Is AI Model Interpretable to Combat with COVID? An Empirical Study on Severity Prediction Task

Black-box nature hinders the deployment of many high-accuracy models in medical diagnosis. Putting one's life in the hands of models that medical researchers do not trust it's irresponsible. However, to understand the mechanism of a new virus, such as COVID-19, machine learning models may catch important symptoms that medical practitioners do not notice due to the surge of infected patients caused by a pandemic. In this work, the interpretation of machine learning models reveals a high CRP corresponds to severe infection, and severe patients usually go through a cardiac injury, which is consistent with medical knowledge. Additionally, through the interpretation of machine learning models, we find phlegm and diarrhea are two important symptoms, without which indicate a high risk of turning severe. These two symptoms are not recognized at the early stage of the outbreak, but later our findings are corroborated by autopsies of COVID-19 patients. And we find patients with a high NTproBNP have a significantly increased risk of death which does not receive much attention initially but proves true by the following-up study. Thus, we suggest interpreting machine learning models can offer help to understanding a new virus at the early stage of an outbreak.

[1]  Stefan Schaal,et al.  From Isolation to Cooperation: An Alternative View of a System of Experts , 1995, NIPS.

[2]  Prognostic value of NT-proBNP in patients with severe COVID-19 , 2020, Respiratory Research.

[3]  Johannes Gehrke,et al.  Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30-day Readmission , 2015, KDD.

[4]  Wojciech Samek,et al.  Methods for interpreting and understanding deep neural networks , 2017, Digit. Signal Process..

[5]  Dongyin Zhang,et al.  Prognostic value of NT-proBNP in patients with severe COVID-19 , 2020, Respiratory Research.

[6]  Been Kim,et al.  Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.

[7]  Daniel W. Apley,et al.  Visualizing the effects of predictor variables in black box supervised learning models , 2016, Journal of the Royal Statistical Society: Series B (Statistical Methodology).

[8]  L. S. Shapley,et al.  17. A Value for n-Person Games , 1953 .

[9]  Amit Dhurandhar,et al.  One Explanation Does Not Fit All: A Toolkit and Taxonomy of AI Explainability Techniques , 2019, ArXiv.

[10]  Sushmita Mitra,et al.  Deep Learning for Screening COVID-19 using Chest X-Ray Images , 2020, 2020 IEEE Symposium Series on Computational Intelligence (SSCI).

[11]  Tommi S. Jaakkola,et al.  Towards Robust Interpretability with Self-Explaining Neural Networks , 2018, NeurIPS.

[12]  Scott Lundberg,et al.  A Unified Approach to Interpreting Model Predictions , 2017, NIPS.

[13]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[14]  Chao Fan,et al.  DeepCOVIDNet: An Interpretable Deep Learning Model for Predictive Surveillance of COVID-19 Using Heterogeneous Features and Their Interactions , 2020, IEEE Access.

[15]  Carlos Guestrin,et al.  Model-Agnostic Interpretability of Machine Learning , 2016, ArXiv.

[16]  Emil Pitkin,et al.  Peeking Inside the Black Box: Visualizing Statistical Learning With Plots of Individual Conditional Expectation , 2013, 1309.6392.

[17]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[18]  T. Singhal A Review of Coronavirus Disease-2019 (COVID-19) , 2020, The Indian Journal of Pediatrics.

[19]  Jaime S. Cardoso,et al.  Machine Learning Interpretability: A Survey on Methods and Metrics , 2019, Electronics.

[20]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[21]  Christian Riess,et al.  A Gentle Introduction to Deep Learning in Medical Image Processing , 2018, Zeitschrift fur medizinische Physik.

[22]  Amit Dhurandhar,et al.  Generating Contrastive Explanations with Monotonic Attribute Functions , 2019, ArXiv.

[23]  Cynthia Rudin,et al.  All Models are Wrong, but Many are Useful: Learning a Variable's Importance by Studying an Entire Class of Prediction Models Simultaneously , 2019, J. Mach. Learn. Res..

[24]  Jonathon Shlens,et al.  Explaining and Harnessing Adversarial Examples , 2014, ICLR.

[25]  L. Mombaerts,et al.  An interpretable mortality prediction model for COVID-19 patients , 2020, Nature Machine Intelligence.

[26]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[27]  Carlos Guestrin,et al.  Anchors: High-Precision Model-Agnostic Explanations , 2018, AAAI.

[28]  L. Wang,et al.  C-reactive protein levels in the early stage of COVID-19 , 2020, Médecine et Maladies Infectieuses.

[29]  Yunming Ye,et al.  DeepFM: A Factorization-Machine based Neural Network for CTR Prediction , 2017, IJCAI.

[30]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[31]  Eri Matsuyama,et al.  A Deep Learning Interpretable Model for Novel Coronavirus Disease (COVID-19) Screening with Chest CT Images , 2020, Journal of Biomedical Science and Engineering.