Automatic structuring of radiology reports: harbinger of a second information revolution in radiology.

Over the past 2 decades, rapid technologic advances have enabled radiologists to acquire, transmit, process, display, and store multidimensional digital images. Despite the steady advances in the way radiologists manage images, similar technologic innovations have not yet affected how radiologists communicate with their colleagues. This discrepancy is particularly striking because the information in radiology reports is just as amenable to computer-based storage, processing, and display as pixels, voxels, and images. In this issue of the journal, Dr Hripcsak and colleagues (1) provide a welcome early look at how information technology will shape the future of radiology reporting (1). They have developed a system that uses natural language processing (NLP) to dissect and structure meaty clinical prose into small digital morsels, each containing a unique medical concept suitable for processing by a computer. Because conventional narrative radiology reports served as early stimuli to refine and improve their NLP system, rigorous evaluations have shown that its accuracy in extracting information from radiology reports is comparable with that of human experts. Their report details an ambitious and impressive evaluation of the NLP system on a huge database of clinical radiology reports encompassing a quarter of a million patients. The authors tasked their system with creating a semantic structure for nearly 900,000 chest radiography reports, which were dictated by radiologists at Columbia-Presbyterian Medical Center over a decade of patient care. An ingenious method was used to externally validate the contents of the structured reports. The frequency and co-occurrence of a variety of clinical conditions was computed from the structured reports and compared with preselected external benchmarks, including peer-reviewed medical literature, regional crime statistics, and financial coding results. These comparisons demonstrate that the structured reports accurately reflected the 3:2 right-to-left ratio of lung cancer, the association of pleural effusions with other clinical conditions, and the decreasing incidence of violent crime in their catchment area. Although not externally validated, their study also produced a fascinating portrait of the semantic contents of radiology reports. For example, their Table 1 provides a unique descriptive look at the occurrence of a variety of clinical conditions in chest radiography reports that ranged from tuberculosis to rib fracture. A separate analysis quantified the radiologist’s impression that a substantial fraction of reports are normal—30% of chest radiography reports in their analysis. The accuracy was comparable with that of expert human coders—achieving 81% sensitivity and 99% specificity, confirming the results of Hripcsak and colleagues’ (1) previous studies. Furthermore, the authors showed that their software outperformed human financial discharge coders in recognizing certain important clinical conditions, such as pneumothorax. Taken together, their results provide a rare look at the contents of radiology reports and illustrate the feasibility of NLP methods to accurately transform databases containing narrative radiology reports into rich sources of information for radiology practices.

[1]  Timothy M. Franz,et al.  Enhancement of clinicians' diagnostic reasoning by computer-based consultation: a multisite study of 2 systems. , 1999, JAMA.

[2]  R A Greenes,et al.  Evaluation of UltraSTAR: performance of a collaborative structured data entry system. , 1994, Proceedings. Symposium on Computer Applications in Medical Care.

[3]  N L Jain,et al.  Respiratory Isolation of Tuberculosis Patients Using Clinical Guidelines and an Automated Clinical Decision Support System , 1998, Infection Control & Hospital Epidemiology.

[4]  Carol Friedman,et al.  Automating SNOMED coding using medical language understanding: a feasibility study , 2001, AMIA.

[5]  George Hripcsak,et al.  Coding Neuroradiology Reports for the Northern Manhattan Stroke Study: A Comparison of Natural Language Processing and Manual Review , 2000, Comput. Biomed. Res..

[6]  C. Langlotz Overcoming barriers to outcomes research on imaging: lessons from an abstract decision model. , 1999, Academic Radiology.

[7]  J. Cimino Desiderata for Controlled Medical Vocabularies in the Twenty-First Century , 1998, Methods of Information in Medicine.

[8]  D. Fardon,et al.  Nomenclature and classification of lumbar disc pathology. Recommendations of the Combined task Forces of the North American Spine Society, American Society of Spine Radiology, and American Society of Neuroradiology. , 2001, Spine.

[9]  A Hanbidge,et al.  Radiology reports: examining radiologist and clinician preferences regarding style and content. , 2001, AJR. American journal of roentgenology.

[10]  Ross D. Shachter,et al.  A Bayesian network for mammography , 2000, AMIA.

[11]  S E Seltzer,et al.  Expediting the turnaround of radiology reports: use of total quality management to facilitate radiologists' report signing. , 1994, AJR. American journal of roentgenology.

[12]  J. Austin,et al.  Glossary of terms for CT of the lungs: recommendations of the Nomenclature Committee of the Fleischner Society. , 1996, Radiology.

[13]  Christoph Wick,et al.  Augmented Reality Simulator for Training in Two-Dimensional Echocardiography , 2000, Comput. Biomed. Res..

[14]  B L Holman,et al.  Medical impact of unedited preliminary radiology reports. , 1994, Radiology.

[15]  J. Austin,et al.  Use of natural language processing to translate clinical information from a database of 889,921 chest radiographic reports. , 2002, Radiology.

[16]  How medical professionals evaluate expressions of probability. , 1987, The New England journal of medicine.

[17]  Stubbs Dm Information content and clarity of radiologists' reports for chest radiography. , 1997 .