Time series models with recurrent neural networks (RNNs) can have high accuracy but are unfortunately difficult to interpret as a result of feature-interactions, temporal-interactions, and nonlinear transformations. Interpretability is important in domains like healthcare where constructing models that provide insight into the relationships they have learned are required to validate and trust model predictions. We want accurate time series models that are interpretable where users can understand the contribution of individual input features. We present the Interpretable-RNN (I-RNN) that balances model complexity and accuracy by forcing the relationship between variables in the model to be additive. Interactions are restricted between hidden states of the RNN and additively combined at the final step. The I-RNN architecture specifically captures the unique characteristics of clinical time series, which are unevenly sampled in time, asynchronously acquired, and have missing data. Importantly, the hidden state activations represent feature coefficients that correlate with the prediction target and can be visualized as risk curves that capture the global relationship between individual input features and the outcome. We evaluate the I-RNN model on the Physionet 2012 Challenge dataset to predict in-hospital mortality, and on a real-world clinical decision support task: predicting hemodynamic interventions in the intensive care unit. I-RNN provides explanations in the form of global and local feature importances comparable to highly intelligible models like decision trees trained on hand-engineered features while significantly outperforming them. Additionally, I-RNN remains intelligible while providing accuracy comparable to state-of-the-art decay-based and interpolation-based recurrent time series models. The experimental results on real-world clinical datasets refute the myth that there is a tradeoff between accuracy and interpretability.
[1]
Shamim Nemati,et al.
Early Prediction of Sepsis From Clinical Data: The PhysioNet/Computing in Cardiology Challenge 2019
,
2019,
2019 Computing in Cardiology (CinC).
[2]
Johannes Gehrke,et al.
Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30-day Readmission
,
2015,
KDD.
[3]
Alistair E. W. Johnson,et al.
Patient specific predictions in the intensive care unit using a Bayesian ensemble
,
2012,
2012 Computing in Cardiology.
[4]
Alun D. Preece,et al.
Interpretability of deep learning models: A survey of results
,
2017,
2017 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computed, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation (SmartWorld/SCALCOM/UIC/ATC/CBDCom/IOP/SCI).
[5]
D. Bates,et al.
Development and Validation of a Deep Learning Algorithm for Mortality Prediction in Selecting Patients With Dementia for Earlier Palliative Care Interventions
,
2019,
JAMA network open.
[6]
Gerardo Flores,et al.
Modelling EHR timeseries by restricting feature interaction
,
2019,
ArXiv.
[7]
Nino Antulov-Fantulin,et al.
Exploring Interpretable LSTM Neural Networks over Multi-Variable Data
,
2019,
ICML.
[8]
Jenna Wiens,et al.
Leveraging Clinical Time-Series Data for Prediction: A Cautionary Tale
,
2018,
AMIA.
[9]
Gerardo Flores,et al.
Explaining an increase in predicted risk for clinical alerts
,
2019,
CHIL.