Incorporating Risk Factor Embeddings in Pre-trained Transformers Improves Sentiment Prediction in Psychiatric Discharge Summaries

Reducing rates of early hospital readmission has been recognized and identified as a key to improve quality of care and reduce costs. There are a number of risk factors that have been hypothesized to be important for understanding re-admission risk, including such factors as problems with substance abuse, ability to maintain work, relations with family. In this work, we develop RoBERTa-based models to predict the sentiment of sentences describing readmission risk factors in discharge summaries of patients with psychosis. We improve substantially on previous results by a scheme that shares information across risk factors while also allowing the model to learn risk factor-specific information.

[1]  James Pustejovsky,et al.  Distinguishing Clinical Sentiment: The Importance of Domain Adaptation in Psychiatric Patient Health Records , 2019, Proceedings of the 2nd Clinical Natural Language Processing Workshop.

[2]  C. Steiner,et al.  Conditions With the Largest Number of Adult Hospital Readmissions by Payer, 2011 , 2006 .

[3]  M. Knapp,et al.  Cost of schizophrenia in England. , 2007, The journal of mental health policy and economics.

[4]  The economic burden of schizophrenia in the United States in 2002. , 2005 .

[5]  R'emi Louf,et al.  HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.

[6]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[7]  E. Rackow Rehospitalizations among patients in the Medicare fee-for-service program. , 2009, The New England journal of medicine.

[8]  Jimmy J. Lin,et al.  Simple BERT Models for Relation Extraction and Semantic Role Labeling , 2019, ArXiv.

[9]  James Pustejovsky,et al.  Analysis of risk factor domains in psychosis patient health records , 2018, Journal of Biomedical Semantics.

[10]  Wiley Interscience The journal of mental health policy and economics , 1998 .

[11]  Wes McKinney,et al.  Data Structures for Statistical Computing in Python , 2010, SciPy.

[12]  Omer Levy,et al.  RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.

[13]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[14]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..