Learning to Select the Best Forecasting Tasks for Clinical Outcome Prediction

The paradigm of ‘pretraining’ from a set of relevant auxiliary tasks and then ‘finetuning’ on a target task has been successfully applied in many different domains. However, when the auxiliary tasks are abundant, with complex relationships to the target task, using domain knowledge or searching over all possible pretraining setups is inefficient and suboptimal. To address this challenge, we propose a method to automatically select from a large set of auxiliary tasks, which yields a representation most useful to the target task. In particular, we develop an efficient algorithm that uses automatic auxiliary task selection within a nested-loop metalearning process. We have applied this algorithm to the task of clinical outcome predictions in electronic medical records, learning from a large number of selfsupervised tasks related to forecasting patient trajectories. Experiments on a real clinical dataset demonstrate the superior predictive performance of our method compared to direct supervised learning, naive pretraining and simple multitask learning, in particular in low-data scenarios when the primary task has very few examples. With detailed ablation analysis, we further show that the selection rules are interpretable and able to generalize to unseen target tasks with new data.

[1]  M. Capuzzo,et al.  Validation of severity scoring systems SAPS II and APACHE II in a single-center population , 2000, Intensive Care Medicine.

[2]  Alan E Jones,et al.  The Sequential Organ Failure Assessment score for predicting outcome in patients with severe sepsis and evidence of hypoperfusion at the time of emergency department presentation* , 2009, Critical care medicine.

[3]  Quoc V. Le,et al.  Semi-supervised Sequence Learning , 2015, NIPS.

[4]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[5]  Peter Szolovits,et al.  MIMIC-III, a freely accessible critical care database , 2016, Scientific Data.

[6]  Ole Winther,et al.  Sequential Neural Models with Stochastic Layers , 2016, NIPS.

[7]  Ole Winther,et al.  A Disentangled Recognition and Nonlinear Dynamics Model for Unsupervised Learning , 2017, NIPS.

[8]  Hugo Larochelle,et al.  Optimization as a Model for Few-Shot Learning , 2016, ICLR.

[9]  Uri Shalit,et al.  Structured Inference Networks for Nonlinear State Space Models , 2016, AAAI.

[10]  Maximilian Karl,et al.  Deep Variational Bayes Filters: Unsupervised Learning of State Space Models from Raw Data , 2016, ICLR.

[11]  Andrew Zisserman,et al.  Multi-task Self-Supervised Visual Learning , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[12]  Yu Zhang,et al.  A Survey on Multi-Task Learning , 2017, IEEE Transactions on Knowledge and Data Engineering.

[13]  Sergey Levine,et al.  Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.

[14]  Paolo Frasconi,et al.  Forward and Reverse Gradient-Based Hyperparameter Optimization , 2017, ICML.

[15]  Paolo Frasconi,et al.  Bilevel Programming for Hyperparameter Optimization and Meta-Learning , 2018, ICML.

[16]  Jeffrey Dean,et al.  Scalable and accurate deep learning with electronic health records , 2018, npj Digital Medicine.

[17]  Vasa Curcin,et al.  Possible Sources of Bias in Primary Care Electronic Health Record Data Use and Reuse , 2018, Journal of medical Internet research.

[18]  Gunnar Rätsch,et al.  Improving Clinical Predictions through Unsupervised Time Series Representation Learning , 2018, ArXiv.

[19]  M. Howell,et al.  Ensuring Fairness in Machine Learning to Advance Health Equity , 2018, Annals of Internal Medicine.

[20]  Joshua Achiam,et al.  On First-Order Meta-Learning Algorithms , 2018, ArXiv.

[21]  Oriol Vinyals,et al.  Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.

[22]  Yu Zhang,et al.  Learning to Multitask , 2018, NeurIPS.

[23]  Jimeng Sun,et al.  Opportunities and challenges in developing deep learning models using electronic health records data: a systematic review , 2018, J. Am. Medical Informatics Assoc..

[24]  Qingfu Zhang,et al.  Pareto Multi-Task Learning , 2019, NeurIPS.

[25]  Mihaela van der Schaar,et al.  Attentive State-Space Modeling of Disease Progression , 2019, NeurIPS.

[26]  Yoshua Bengio,et al.  Unsupervised State Representation Learning in Atari , 2019, NeurIPS.

[27]  Sergey Levine,et al.  Unsupervised Learning via Meta-Learning , 2018, ICLR.

[28]  Allan Jabri,et al.  Unsupervised Curricula for Visual Meta-Reinforcement Learning , 2019, NeurIPS.

[29]  Parisa Rashidi,et al.  DeepSOFA: A Continuous Acuity Score for Critically Ill Patients using Clinically Interpretable Deep Learning , 2018, Scientific Reports.

[30]  Suman V. Ravuri,et al.  A Clinically Applicable Approach to Continuous Prediction of Future Acute Kidney Injury , 2019, Nature.

[31]  Jiayu Zhou,et al.  MetaPred: Meta-Learning for Clinical Risk Prediction with Limited Patient Electronic Health Records , 2019, KDD.

[32]  Jascha Sohl-Dickstein,et al.  Meta-Learning Update Rules for Unsupervised Representation Learning , 2018, ICLR.

[33]  Yuan Xue,et al.  Deep Physiological State Space Model for Clinical Forecasting , 2019, ArXiv.

[34]  Sebastian Nowozin,et al.  Fast and Flexible Multi-Task Classification Using Conditional Neural Adaptive Processes , 2019, NeurIPS.

[35]  Ramakanth Pasunuru,et al.  AutoSeM: Automatic Task Selection and Mixing in Multi-Task Learning , 2019, NAACL.

[36]  Karsten M. Borgwardt,et al.  Early prediction of circulatory failure in the intensive care unit using machine learning , 2020, Nature Medicine.

[37]  Gholamreza Haffari,et al.  Learning to Multi-Task Learn for Better Neural Machine Translation , 2020, ArXiv.

[38]  Timothy M. Hospedales,et al.  Meta-Learning in Neural Networks: A Survey , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.