Effects of Computer-based Clinical Decision Support Systems on Clinician Performance and Patient Outcome: A Critical Appraisal of Research

The application of artificial intelligence and other computing and information science techniques to the field of health care has resulted in the development of computer-based clinical decision support systems (CDSSs), sometimes called, generically, expert systems. Although no consensus has been achieved on the definition of a CDSS, Wyatt and Spiegelhalter [1] have defined medical decision aids as active knowledge systems which use two or more items of patient data to generate case-specific advice, thus capturing the main attributes of these systems in a simple statement. Much has been written about the theoretical and technical aspects of CDSSs and their reliability, validity, and acceptability. Wyatt and Spiegelhalter [2] described a systematic approach to laboratory and field testing of CDSSs, suggesting that the final stages should include evaluation of effects on the health care process and on patient outcomes. This overview focuses on studies of the final and most clinically important stages of evaluation and examines controlled trials designed to measure the effects of CDSSs on clinician performance and patient outcomes. Methods Study Identification Previously published reviews of CDSSs were identified using a MEDLINE search from January 1983 through February 1992 and through a manual search of textbooks and conference proceedings in the areas of artificial intelligence and computer applications in medicine. Original studies were identified by an all-language MEDLINE search of articles published from 1974 (the year of publication of the first article to evaluate the effect of a CDSS on clinician performance [3]) to February 1992 (search terms available on request). Studies were also identified through an update of a previous review on computer-aided quality assurance [4]; through an EMBASE (Excerpta Medica) search for the same time period; through an INSPEC (International Information Service for the Physics and Engineering Communities) search; through review of citations in the articles from electronic searches and a search forward on three citations [5-7], one each from the areas of dose determination, diagnosis, and quality assurance, using SCISEARCH; through articles on related topics collected by the Health Information Research Unit of McMaster University, including a regularly updated bibliography of studies of continuing education [8]; and by scanning the Proceedings of the Symposium on Computer Applications in Medical Care, 1989 through 1991. After a set of relevant publications was selected for inclusion in the overview, a list of their titles was sent to corresponding authors and experts in medical informatics with a request for information about any additional published or unpublished studies. Study Selection We selected studies for review if 1) the population of interest was composed of clinicians [physicians, nurses, dentists, and so on] in practice or training; 2) the intervention was a computer-based CDSS evaluated in a clinical setting; 3) the outcomes assessed were clinician performance, a measure of the process of care, or patient outcomes, including any aspect of patient well-being; and 4) the type of evidence was limited to prospective studies with a contemporaneous control group where patient care with a CDSS was compared with patient care without one. Crossover studies were included. A CDSS was defined as follows: computer software using a knowledge base designed for use by a clinician involved in patient care as a direct aid to clinical decision making. Characteristics of an individual patient were matched to information in the knowledge base. Patient-specific information in the form of assessments (management options or probabilities) or recommendations were presented to the clinician. Lists of citations, with abstracts and indexing terms if available, were assessed by two of the authors, one with a background in expert system development and one with experience in critical appraisal of applied medical research. Each rater indicated whether the citation was potentially relevant (that is, appeared to meet the selection criteria, applied loosely at this stage), was clearly not relevant, or gave insufficient information to make a judgment. The latter usually lacked abstracts; the full texts of these articles were examined by one of the raters in the library and rated as potentially relevant or definitely irrelevant. A copy was made of the full article for citations judged to be potentially relevant by either rater and then rated independently by both raters according to the four criteria listed above. To be included in the overview, a study was required to meet all the selection criteria. A third rater independently performed the same detailed relevance rating if the original raters disagreed or were uncertain about the study's relevance to the overview. In these cases, a study was included if two of the three raters deemed it relevant. Study Evaluation Two reviewers independently rated the studies selected for the overview on each of five potential sources of bias (see the Appendix). Disagreements were handled in the same way as those encountered during study selection. Additional information about study design was solicited from authors if necessary. Options under each item were assigned a score of 2, 1, or 0 points. Scores were summed for the five items to achieve an overall score; 10 was the highest possible score. The first rating option under each item indicated that appropriate steps were taken to minimize bias in the study results; the second option indicated that although efforts were made to reduce bias, they may not have been adequate; and the third option indicated a study in which a definite potential for bias existed or in which insufficient information was given to assure the reader that no bias was present. Item three (unit of allocation) was a measure of potential contamination specific to evaluations in which interventions are applied to clinicians and end points are measured for patients. When individual patients are allocated to intervention and control groups, clinicians may treat patients with or without the aid of the CDSS. Because only a portion of the patients in a given study may have been treated with the aid of a CDSS, knowledge gained from the CDSS may be applied to control patients, leading to an underestimate of the system's effect. In some cases, the CDSS may be accessible during assessment of control patients and could be used in their care. Similar problems can occur when individual clinicians are the unit of allocation. The presence of a CDSS or of colleagues who are using a CDSS in the clinic may influence the treatment given by control clinicians. Item five, concerning follow-up rates, was not applicable for several studies of computer-aided quality assurance in which all patients in the practice were randomly assigned to be evaluated with the aid of a CDSS or to be control patients but in which data were generated only for those coming to the clinic for a visit during the study period. These studies were scored on the first four criteria and the total was prorated so that the maximum possible score was 10 points. Data Extraction Data concerning setting, clinicians and patients, interventions, and outcomes were extracted from each article by one investigator using a structured data collection form. Data extraction was verified by a second rater, and corrections were made where necessary. A letter was sent to the corresponding author for each study to request missing data and verification of the description of the study setting, subjects, CDSS, and the main outcome to be included in the overview. Analysis We began the overview expecting that studies would come from several areas of health care and would evaluate a number of CDSSs developed for different purposes. They would be assessed in several settings using many types of outcomes: continuous, discrete, long-term, and short-term, with some having more direct clinical relevance than others. A single measure of effect, in our judgment, would not be the best expression of our current knowledge of the effects of CDSSs on clinicians and patients, and a formal meta-analysis could be misleading for most applications. We decided, instead, to present summaries of the effect of each type of system for several clinician and patient outcomes separately so that system developers and consumers could focus on the type of system and outcome that would be relevant to their practice. An effect size was calculated for each study reporting a continuous outcome variable by dividing the difference of the means for the intervention and control groups by the pooled standard deviation. The effect size indicated the number of standard deviations by which the CDSS and control groups differed. For dichotomous outcome variables, odds ratios were calculated. An odds ratio whose 95% CI did not include 1 indicated a significant beneficial effect of the CDSS. An odds ratio less than 1 favored the CDSS when the outcome variable of interest was the number of adverse effects, and an odds ratio of greater than 1 favored the CDSS when the outcome variable measured health or clinician performance. Differences in the findings for studies of the same type of CDSS were assessed statistically for heterogeneity using the Homogeneity Q test [9] or the Breslow-Day test [10]. Two crossover studies were treated as parallel designs because the risk for carryover effects from one study period to the next was considered minimal. Results Retrieval of Previous Reviews and Original Studies We identified seven previous reviews that examined experimental evidence of the effects of CDSSs on the process or outcome of care [4, 11-16]. Only one [4] was a systematic overview, and this was restricted to the evidence available by 1986 concerning computer-aided quality assurance, which is only one of several types of CDSSs available. To find reports of orig

[1]  F. T. de Dombal,et al.  Human and Computer-aided Diagnosis of Abdominal Pain: Further Report with Emphasis on Performance of Clinicians , 1974, British medical journal.

[2]  F. Oski,et al.  Impact of a system of computer-assisted diagnosis. Initial evaluation of the hospitalized patient. , 1975, American journal of diseases of children.

[3]  C. Mc Donald,et al.  Use of a computer to detect and respond to clinical events: its effect on clinician behavior. , 1976, Annals of internal medicine.

[4]  C. McDonald Protocol-based computer reminders, the quality of care and the non-perfectability of man. , 1976, The New England journal of medicine.

[5]  S. Oparil,et al.  Treatment of hypertension by computer and physician-a prospective controlled study. , 1977, Journal of chronic diseases.

[6]  E.H. Shortliffe,et al.  Knowledge engineering for medical decision making: A review of computer-based clinical decision aids , 1979, Proceedings of the IEEE.

[7]  J L Rogers,et al.  The Impact of a Computerized Medical Record Summary System on Incidence and Length of Hospitalization , 1979, Medical care.

[8]  C J McDonald,et al.  Physician response to computer reminders. , 1980, JAMA.

[9]  R B D'Agostino,et al.  The usefulness of a predictive instrument to reduce inappropriate admissions to the coronary care unit. , 1980, Annals of internal medicine.

[10]  D W Young Improving the consistency with which investigations are requested. , 1981, Medical informatics = Medecine et informatique.

[11]  P M Wortman,et al.  Medical Information Systems: Assessing Impact in the Areas of Hypertension, Obesity and Renal Disease , 1982, Medical care.

[12]  G. Barnett,et al.  A Computer-Based Monitoring System for Follow-Up of Elevated Blood Pressure , 1983, Medical care.

[13]  T Shinozaki,et al.  Medical Information Management: Improving the Transfer of Research Results to Presurgical Evaluation , 1983, Medical care.

[14]  R W Jelliffe,et al.  Clinical studies with computer-assisted initial lidocaine therapy. , 1984, Archives of internal medicine.

[15]  C. McDonald,et al.  Reminders to physicians from an introspective computer medical record. A two-year randomized trial. , 1984, Annals of internal medicine.

[16]  O. Haring,et al.  Changes in patient attitudes following the implementation of a medical information system. , 1984, QRB. Quality review bulletin.

[17]  R B D'Agostino,et al.  A predictive instrument to improve coronary-care-unit admission practices in acute ischemic heart disease. A prospective multicenter clinical trial. , 1984, The New England journal of medicine.

[18]  T A Pryor,et al.  Application of a computerized medical decision-making process to the problem of digoxin intoxication. , 1984, Journal of the American College of Cardiology.

[19]  L. Hedges,et al.  Statistical Methods for Meta-Analysis , 1987 .

[20]  I McDowell,et al.  Comparison of three methods of recalling patients for influenza vaccination. , 1986, CMAJ : Canadian Medical Association journal = journal de l'Association medicale canadienne.

[21]  C J McDonald,et al.  Delayed Feedback of Physician Performance Versus Immediate Reminders to Perform Preventive Care: Effects on Physician Compliance , 1986, Medical care.

[22]  H. Covvey,et al.  Randomised controlled trial of computer assisted management of hypertension in primary care. , 1986, British medical journal.

[23]  P L Miller,et al.  The evaluation of artificial intelligence systems in medicine. , 1985, Computer methods and programs in biomedicine.

[24]  G Brownbridge,et al.  An interactive computerized protocol for the management of hypertension: effects on the general practitioner's clinical behaviour. , 1986, The Journal of the Royal College of General Practitioners.

[25]  N E Day,et al.  Statistical methods in cancer research. IARC Workshop 25-27 May 1983. , 1987, IARC scientific publications.

[26]  R. Haynes,et al.  Computer-aided quality assurance. A critical appraisal. , 1987, Archives of internal medicine.

[27]  M. Daschbach,et al.  Initiation of warfarin therapy: comparison of physician dosing with computer-assisted dosing. , 1987, Journal of general internal medicine.

[28]  E. Shortliffe Computer programs to support clinical decision making. , 1990, JAMA.

[29]  C J McDonald,et al.  Computer predictions of abnormal test results. Effects on outpatient testing. , 1988, JAMA.

[30]  I McDowell,et al.  Computerized reminders to encourage cervical screening in family practice. , 1989, The Journal of family practice.

[31]  G C Sutton,et al.  Computer‐aided diagnosis: A review , 1989, The British journal of surgery.

[32]  I McDowell,et al.  A Randomized Trial of Computerized Reminders for Blood Pressure Screening in Primary Care , 1989, Medical care.

[33]  J P Ornato,et al.  Computer-assisted optimization of aminophylline therapy in the emergency department. , 1989, The American journal of emergency medicine.

[34]  S. Paterson-Brown,et al.  Modern aids to clinical decision‐making in the acute abdomen , 1990, The British journal of surgery.

[35]  D. Spiegelhalter,et al.  Evaluating medical expert systems: what to test and how? , 1990, Medical informatics = Medecine et informatique.

[36]  J G Mazoué,et al.  Diagnosis without doctors. , 1990, The Journal of medicine and philosophy.

[37]  K Canfield,et al.  Evaluation of UNIS: Urological Nursing Information Systems. , 1991, Proceedings. Symposium on Computer Applications in Medical Care.

[38]  L S Keller,et al.  Effects of computerized nurse careplanning on selected health care effectiveness measures. , 1991, Proceedings. Symposium on Computer Applications in Medical Care.

[39]  D. Spiegelhalter,et al.  Field trials of medical decision-aids: potential problems and solutions. , 1991, Proceedings. Symposium on Computer Applications in Medical Care.

[40]  D. Mungall,et al.  Outpatient management of warfarin therapy: comparison of computer-predicted dosage adjustment to skilled professional care. , 1991, Therapeutic drug monitoring.

[41]  D J Spiegelhalter,et al.  How does computer-aided diagnosis improve the management of acute abdominal pain? , 1992, Annals of the Royal College of Surgeons of England.

[42]  I McDowell,et al.  Use of reminders to increase compliance with tetanus booster vaccination. , 1992, CMAJ : Canadian Medical Association journal = journal de l'Association medicale canadienne.