Generating Medical Reports from Patient-Doctor Conversations Using Sequence-to-Sequence Models

We discuss automatic creation of medical reports from ASR-generated patient-doctor conversational transcripts using an end-to-end neural summarization approach. We explore both recurrent neural network (RNN) and Transformer-based sequence-to-sequence architectures for summarizing medical conversations. We have incorporated enhancements to these architectures, such as the pointer-generator network that facilitates copying parts of the conversations to the reports, and a hierarchical RNN encoder that makes RNN training three times faster with long inputs. A comparison of the relative improvements from the different model architectures over an oracle extractive baseline is provided on a dataset of 800k orthopedic encounters. Consistent with observations in literature for machine translation and related tasks, we find the Transformer models outperform RNN in accuracy, while taking less than half the time to train. Significantly large wins over a strong oracle baseline indicate that sequence-to-sequence modeling is a promising approach for automatic generation of medical reports, in the presence of data at scale.

[1]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[2]  Rikin Patel,et al.  Factors Related to Physician Burnout and Its Consequences: A Review , 2018, Behavioral sciences.

[3]  Tobias Domhan,et al.  How Much Attention Do You Need? A Granular Analysis of Neural Machine Translation Architectures , 2018, ACL.

[4]  Juan C. Quiroz,et al.  Challenges of developing a digital scribe to reduce clinical documentation burden , 2019, npj Digital Medicine.

[5]  Thomas H. Payne,et al.  Using voice to create hospital progress notes: Description of a mobile application and supporting system integrated with a commercial electronic health record , 2018, J. Biomed. Informatics.

[6]  Andrew Gordon Wilson,et al.  Averaging Weights Leads to Wider Optima and Better Generalization , 2018, UAI.

[7]  Anita Burgun-Parenthoine,et al.  Automatic concept extraction from spoken medical reports , 2003, Int. J. Medical Informatics.

[8]  Christine A. Sinsky,et al.  Relationship Between Clerical Burden and Characteristics of the Electronic Environment With Physician Burnout and Professional Satisfaction. , 2016, Mayo Clinic proceedings.

[9]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[10]  Ankur Bapna,et al.  The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation , 2018, ACL.

[11]  Hagen Soltau,et al.  The Medical Scribe: Corpus Development and Model Performance Analyses , 2020, LREC.

[12]  Chin-Yew Lin,et al.  ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.

[13]  Christopher D. Manning,et al.  Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.

[14]  Mirella Lapata,et al.  Don’t Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization , 2018, EMNLP.

[15]  Christopher D. Manning,et al.  Learning to Summarize Radiology Findings , 2018, Louhi@EMNLP.

[16]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[17]  Nan Du,et al.  Extracting Symptoms and their Status from Clinical Conversations , 2019, ACL.

[18]  Gang Li,et al.  Learning to Infer Entities, Properties and their Relations from Clinical Conversations , 2019, EMNLP.

[19]  Evangelos Kontopantelis,et al.  Controlled Interventions to Reduce Burnout in Physicians: A Systematic Review and Meta-analysis , 2017, JAMA Internal Medicine.

[20]  Geoffrey E. Hinton,et al.  Layer Normalization , 2016, ArXiv.

[21]  Ashish Vaswani,et al.  Self-Attention with Relative Position Representations , 2018, NAACL.

[22]  Parth Pathak,et al.  Annotation of a Large Clinical Entity Corpus , 2018, EMNLP.

[23]  Nazli Goharian,et al.  Ontology-Aware Clinical Abstractive Summarization , 2019, SIGIR.

[24]  Bowen Zhou,et al.  Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond , 2016, CoNLL.

[25]  Sandeep Konam,et al.  Medication Regimen Extraction From Clinical Conversations , 2019, ArXiv.

[26]  David Suendermann-Oeft,et al.  An Automated Assistant for Medical Scribes , 2018, INTERSPEECH.

[27]  Christine A. Sinsky,et al.  Allocation of Physician Time in Ambulatory Practice: A Time and Motion Study in 4 Specialties , 2016, Annals of Internal Medicine.

[28]  Y. Nesterov A method for solving the convex programming problem with convergence rate O(1/k^2) , 1983 .

[29]  Jingbo Zhu,et al.  Learning Deep Transformer Models for Machine Translation , 2019, ACL.

[30]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[31]  Gondy Leroy,et al.  Automated Extraction of Diagnostic Criteria From Electronic Health Records for Autism Spectrum Disorders: Development, Evaluation, and Application , 2018, Journal of medical Internet research.

[32]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[33]  Franck Dernoncourt,et al.  A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents , 2018, NAACL.

[34]  Stephan Peitz,et al.  Jointly Learning to Align and Translate with Transformer Models , 2019, EMNLP.

[35]  I. Katz,et al.  Beyond Burnout - Redesigning Care to Restore Meaning and Sanity for Physicians. , 2018, The New England journal of medicine.

[36]  David Suendermann-Oeft,et al.  An automated medical scribe for documenting clinical encounters , 2018, NAACL.