From dictations to clinical reports using machine translation

A typical workflow to document clinical encounters entails dictating a summary, running speech recognition, and post-processing the resulting text into a formatted letter. Post-processing entails a host of transformations including punctuation restoration, truecasing, marking sections and headers, converting dates and numerical expressions, parsing lists, etc. In conventional implementations, most of these tasks are accomplished by individual modules. We introduce a novel holistic approach to post-processing that relies on machine callytranslation. We show how this technique outperforms an alternative conventional system—even learning to correct speech recognition errors during post-processing—while being much simpler to maintain.

[1]  Janice C. Honeyman-Buck,et al.  Managing predefined templates and macros for a departmental speech recognition system using common software , 2001, Journal of Digital Imaging.

[2]  David Suendermann-Oeft,et al.  Automated Preamble Detection in Dictated Medical Reports , 2017, BioNLP.

[3]  Morten Hertzum,et al.  Acceptance of speech recognition by physicians: A survey of expectations, experiences, and social influence , 2009, Int. J. Hum. Comput. Stud..

[4]  Franz Josef Och,et al.  Minimum Error Rate Training in Statistical Machine Translation , 2003, ACL.

[5]  Markus Freitag,et al.  Modeling punctuation prediction as machine translation , 2011, IWSLT.

[6]  David Liu,et al.  Six Characteristics of Effective Structured Reporting and the Inevitable Integration with Speech Recognition , 2005, Journal of Digital Imaging.

[7]  John D. Lafferty,et al.  Cyberpunc: a lightweight punctuation annotation system for speech , 1998, Proceedings of the 1998 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP '98 (Cat. No.98CH36181).

[8]  David Suendermann-Oeft,et al.  Deep Learning for Punctuation Restoration in Medical Reports , 2017, BioNLP.

[9]  Linda Dawson,et al.  A systematic review of speech recognition technology in health care , 2014, BMC Medical Informatics and Decision Making.

[10]  Hermann Ney,et al.  CDER: Efficient MT Evaluation Using Block Movements , 2006, EACL.

[11]  Hermann Ney,et al.  Algorithms for statistical translation of spoken language , 2000, IEEE Trans. Speech Audio Process..

[12]  Hermann Ney,et al.  A Systematic Comparison of Various Statistical Alignment Models , 2003, CL.

[13]  Joshua C Denny,et al.  Generating Clinical Notes for Electronic Health Record Systems , 2010, Applied Clinical Informatics.

[14]  David Suendermann-Oeft,et al.  Medical Speech Recognition: Reaching Parity with Humans , 2017, SPECOM.

[15]  Daniel Marcu,et al.  Statistical Phrase-Based Translation , 2003, NAACL.

[16]  Enrico W. Coiera,et al.  Risks and benefits of speech recognition for clinical documentation: a systematic review , 2016, J. Am. Medical Informatics Assoc..

[17]  Philipp Koehn,et al.  Moses: Open Source Toolkit for Statistical Machine Translation , 2007, ACL.

[18]  Anoop Deoras,et al.  Decoding-time prediction of non-verbalized punctuation , 2008, INTERSPEECH.

[19]  Miles Osborne,et al.  Statistical Machine Translation , 2010, Encyclopedia of Machine Learning and Data Mining.