Continuous Patient State Attention Models

Irregular time-series (ITS) are prevalent in the electronic health records (EHR) as the data is recorded in EHR system as per the clinical guidelines/requirements but not for research and also depends on the patient health status. ITS present challenges in training of machine learning algorithms, which are mostly built on assumption of coherent fixed dimensional feature space. In this paper, we propose a computationally efficient variant of the transformer based on the idea of cross-attention, called Perceiver, for time-series in healthcare. We further develop continuous patient state attention models, using the Perceiver and the transformer to deal with ITS in EHR. The continuous patient state models utilise neural ordinary differential equations to learn the patient health dynamics, i.e., patient health trajectory from the observed irregular time-steps, which enables them to sample any number of time-steps at any time. The performance of the proposed models is evaluated on in-hospital-mortality prediction task on Physionet-2012 challenge and MIMIC-III datasets. The Perceiver model significantly outperforms the baselines and reduces the computational complexity, as compared with the transformer model, without significant loss of performance. The carefully designed experiments to study irregularity in healthcare also show that the continuous patient state models outperform the baselines. The code is publicly released and verified at https://codeocean.com/capsule/4587224

[1]  D. Clifton,et al.  Improving Classification of Tetanus Severity for Patients in Low-Middle Income Countries Wearing ECG Sensors by Using a CNN-Transformer Network , 2022, IEEE Transactions on Biomedical Engineering.

[2]  D. Clifton,et al.  COPER: Continuous Patient State Perceiver , 2022, 2022 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI).

[3]  Heung-Il Suk,et al.  Multi-View Integrative Attention-Based Deep Representation Learning for Irregular Clinical Time-Series Data , 2022, IEEE Journal of Biomedical and Health Informatics.

[4]  Oriol Vinyals,et al.  General-purpose, long-context autoregressive modeling with Perceiver AR , 2022, ICML.

[5]  D. Clifton,et al.  Dynamic Neural Graphs Based Federated Reptile for Semi-Supervised Multi-Tasking in Healthcare Applications , 2021, IEEE Journal of Biomedical and Health Informatics.

[6]  Olivier J. H'enaff,et al.  Perceiver IO: A General Architecture for Structured Inputs & Outputs , 2021, ICLR.

[7]  Chandan K. Reddy,et al.  Self-Supervised Transformer for Sparse and Irregularly Sampled Multivariate Clinical Time-Series , 2021, ACM Trans. Knowl. Discov. Data.

[8]  Andrew Zisserman,et al.  Perceiver: General Perception with Iterative Attention , 2021, ICML.

[9]  Stephanie L. Hyland,et al.  Temporal pointwise convolutional networks for length of stay prediction in the intensive care unit , 2020, CHIL.

[10]  Satya Narayan Shukla,et al.  Multi-Time Attention Networks for Irregularly Sampled Time Series , 2020, ICLR.

[11]  Shenda Hong,et al.  A Review of Deep Learning Methods for Irregularly Sampled Medical Time Series Data , 2020, ArXiv.

[12]  Baoyao Yang,et al.  DATA-GRU: Dual-Attention Time-Aware Gated Recurrent Unit for Irregular Multivariate Time Series , 2020, AAAI.

[13]  Hagen Malberg,et al.  Predicting sepsis with a recurrent neural network using the MIMIC III database , 2019, Comput. Biol. Medicine.

[14]  Aram Galstyan,et al.  Multitask learning and benchmarking with clinical time series data , 2017, Scientific Data.

[15]  Jinsung Yoon,et al.  Estimating Missing Data in Temporal Data Streams Using Multi-Directional Recurrent Neural Networks , 2017, IEEE Transactions on Biomedical Engineering.

[16]  David Duvenaud,et al.  Latent Ordinary Differential Equations for Irregularly-Sampled Time Series , 2019, NeurIPS.

[17]  David Duvenaud,et al.  Neural Ordinary Differential Equations , 2018, NeurIPS.

[18]  Vladlen Koltun,et al.  An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling , 2018, ArXiv.

[19]  Andreas Spanias,et al.  Attend and Diagnose: Clinical Time Series Analysis using Attention Models , 2017, AAAI.

[20]  Parisa Rashidi,et al.  Deep EHR: A Survey of Recent Advances in Deep Learning Techniques for Electronic Health Record (EHR) Analysis , 2017, IEEE Journal of Biomedical and Health Informatics.

[21]  Luca Antiga,et al.  Automatic differentiation in PyTorch , 2017 .

[22]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[23]  Peter Szolovits,et al.  MIMIC-III, a freely accessible critical care database , 2016, Scientific Data.

[24]  Zoubin Ghahramani,et al.  Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.

[25]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[26]  Benjamin M. Marlin,et al.  Unsupervised pattern discovery in electronic health care data using probabilistic clustering models , 2012, IHI '12.

[27]  Jeffrey M. Hausdorff,et al.  Physionet: Components of a New Research Resource for Complex Physiologic Signals". Circu-lation Vol , 2000 .

[28]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.