Attentive State-Space Modeling of Disease Progression

Models of disease progression are instrumental for predicting patient outcomes and understanding disease dynamics. Existing models provide the patient with pragmatic (supervised) predictions of risk, but do not provide the clinician with intelligible (unsupervised) representations of disease pathophysiology. In this paper, we develop the attentive state-space model, a deep probabilistic model that learns accurate and interpretable structured representations for disease trajectories. Unlike Markovian state-space models, in which the dynamics are memoryless, our model uses an attention mechanism to create "memoryful" dynamics, whereby attention weights determine the dependence of future disease states on past medical history. To learn the model parameters from medical records, we develop an infer ence algorithm that simultaneously learns a compiled inference network and the model parameters, leveraging the attentive state-space representation to construct a "Rao-Blackwellized" variational approximation of the posterior state distribution. Experiments on data from the UK Cystic Fibrosis registry show that our model demonstrates superior predictive accuracy and provides insights into the progression of chronic disease.

[1]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[2]  Eric J Topol,et al.  High-performance medicine: the convergence of human and artificial intelligence , 2019, Nature Medicine.

[3]  Jimeng Sun,et al.  RETAIN: An Interpretable Predictive Model for Healthcare using Reverse Time Attention Mechanism , 2016, NIPS.

[4]  Le Song,et al.  Recurrent Hidden Semi-Markov Model , 2017, ICLR.

[5]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[6]  Mihaela van der Schaar,et al.  Disease-Atlas: Navigating Disease Trajectories with Deep Learning , 2018, ArXiv.

[7]  Le Song,et al.  Efficient Learning of Continuous-Time Hidden Markov Models for Disease Progression , 2015, NIPS.

[8]  Ran El-Yaniv,et al.  On Prediction Using Variable Order Markov Models , 2004, J. Artif. Intell. Res..

[9]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[10]  Maximilian Karl,et al.  Deep Variational Bayes Filters: Unsupervised Learning of State Space Models from Raw Data , 2016, ICLR.

[11]  Alexander J. Smola,et al.  State Space LSTM Models with Particle MCMC Inference , 2017, ArXiv.

[12]  Mihaela van der Schaar,et al.  Learning from Clinical Judgments: Semi-Markov-Modulated Marked Hawkes Processes for Risk Prognosis , 2017, ICML.

[13]  Sean Gerrish,et al.  Black Box Variational Inference , 2013, AISTATS.

[14]  Jimeng Sun,et al.  RetainVis: Visual Analytics with Interpretable and Interactive Recurrent Neural Networks on Electronic Medical Records , 2018, IEEE Transactions on Visualization and Computer Graphics.

[15]  P. Green,et al.  Hidden Markov Models and Disease Mapping , 2002 .

[16]  C. Merlo,et al.  Cystic fibrosis lung transplantation , 2011, Current opinion in pulmonary medicine.

[17]  Ryan P. Adams,et al.  Composing graphical models with neural networks for structured representations and fast inference , 2016, NIPS.

[18]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[19]  Mihaela van der Schaar,et al.  A Hidden Absorbing Semi-Markov Model for Informatively Censored Temporal Data: Learning and Inference , 2016, J. Mach. Learn. Res..

[20]  E. Neilson,et al.  Chronic kidney disease progression. , 2006, Journal of the American Society of Nephrology : JASN.

[21]  Xiang Wang,et al.  Unsupervised learning of disease progression models , 2014, KDD.

[22]  Charles Elkan,et al.  Learning to Diagnose with LSTM Recurrent Neural Networks , 2015, ICLR.

[23]  Gretchen A. Piatt,et al.  Patients with Complex Chronic Diseases: Perspectives on Supporting Self-Management , 2007, Journal of General Internal Medicine.

[24]  D. Sanders,et al.  Return of FEV1 after pulmonary exacerbation in children with cystic fibrosis , 2010, Pediatric pulmonology.

[25]  Ole Winther,et al.  Sequential Neural Models with Stochastic Layers , 2016, NIPS.

[26]  Matthias W. Seeger,et al.  Deep State Space Models for Time Series Forecasting , 2018, NeurIPS.

[27]  B. Starfield,et al.  Defining Comorbidity: Implications for Understanding Health and Health Services , 2009, The Annals of Family Medicine.

[28]  John Pestian,et al.  Phenotypes of Rapid Cystic Fibrosis Lung Disease Progression during Adolescence and Young Adulthood , 2017, American journal of respiratory and critical care medicine.

[29]  Fenglong Ma,et al.  Dipole: Diagnosis Prediction in Healthcare via Attention-based Bidirectional Recurrent Neural Networks , 2017, KDD.

[30]  Christian Osendorfer,et al.  Learning Stochastic Recurrent Networks , 2014, NIPS 2014.

[31]  Karol Gregor,et al.  Neural Variational Inference and Learning in Belief Networks , 2014, ICML.

[32]  Yoshua Bengio,et al.  A Recurrent Latent Variable Model for Sequential Data , 2015, NIPS.

[33]  Walter F. Stewart,et al.  Doctor AI: Predicting Clinical Events via Recurrent Neural Networks , 2015, MLHC.

[34]  Y. Shtarkov,et al.  The context-tree weighting method: basic properties , 1995, IEEE Trans. Inf. Theory.

[35]  Mihaela van der Schaar,et al.  Estimating Counterfactual Treatment Outcomes over Time Through Adversarially Balanced Representations , 2020, ICLR.

[36]  Max Welling,et al.  Auto-Encoding Variational Bayes , 2013, ICLR.

[37]  D. Blumenthal,et al.  The "meaningful use" regulation for electronic health records. , 2010, The New England journal of medicine.

[38]  Pieter Abbeel,et al.  Gradient Estimation Using Stochastic Computation Graphs , 2015, NIPS.

[39]  Uri Shalit,et al.  Structured Inference Networks for Nonlinear State Space Models , 2016, AAAI.