IN THE EARLY 1890S, DR WILLIAM HALSTED DEVELOPED radical mastectomy for breast cancer. Surgeons performed the Halsted procedure for more than 80 years even though there was little systematic evidence for its success. Then a new breed of scholars subjected the procedure to formal methods of evaluation unknown to Halsted. The methods—randomized controlled trials (RCTs) principal among them—led to a surprise: radical mastectomy had no advantage over simpler forms of treatment. This is but 1 example of the hard-won victory of evidence over belief in medicine. The pioneers of the formal evaluation of medical practices raised questions that traditional practitioners did not welcome. But in time, formal evaluation prevailed. The pioneers developed a hierarchy of evidentiary rigor relating the design of a study to the confidence that could be placed in the findings, from the lowly, nearly valueless anecdote to the royalty of evidence, the RCT. Concurrently, a similar story of hard-won learning unfolded in the so-called quality movement. Scholars illuminated the scale and types of defects in the processes of care and the outcomes, including high rates of unscientific care, inappropriate care, geographic variations in practice, latent disagreements among specialists, and oftenunrecognized medical injury to patients. Like the pioneers of evidence-based medicine, students of medical quality were at first largely ignored, but no longer. In 1999 and 2001, the Institute of Medicine published 2 landmark reports on the evidence for quality failures and called urgently for redesign of care systems to achieve improvements. The story could end here happily with 2 great streams of endeavor merging into a framework for conjoint action: improving clinical evidence and improving the process of care. Instead, the 2 endeavors are often in unhappy tension. Neither disputes that progress toward health care’s main goal, the relief of illness and pain, requires research of many kinds: basic, clinical, systems, epidemiologic. The disagreement centers on epistemology—ways to get at “truth” and how those ways should vary depending on the knowledge sought. Individuals most involved in day-to-day improvement work fear that if “evidence” is too narrowly defined and the approach to gathering evidence too severely constrained, progress may be the victim. For example, the RCT is a powerful, perhaps unequaled, research design to explore the efficacy of conceptually neat components of clinical practice—tests, drugs, and procedures. For other crucially important learning purposes, however, it serves less well. Recent controversies about the evaluation of rapid response teams provide a case in point. These controversies show the importance of adjusting research methods to fit research questions. Although only 10% to 15% of inpatients resuscitated outside intensive care units survive to hospital discharge, early warning signs are present in a large percentage of patients who ultimately experience cardiac arrest. Rapid response team systems bring expert clinicians to the bedsides of deteriorating patients before arrest occurs. In the mid 1990s, based largely on reports from Australian investigators, the Institute for Healthcare Improvement and others began introducing the concept to willing hospitals. Local experience strongly suggested that these systems often, although not always, were associated with improved outcomes, including reduced anxiety among nursing staff; increased interdisciplinary teamwork; decreased cardiac arrests outside of intensive care units; and, in some cases, declines in mortality. The evidence base took a turn in June 2005 with the publication of the Medical Early Response Intervention and Therapy (MERIT) Study, a cluster randomized prospective trial that claimed to find no beneficial effect of these teams on several primary outcomes. Controversy has continued since then regarding the scientific evidence for rapid response systems. In fact, the MERIT trial was not negative; it was inconclusive. The study team encountered an array of serious problems in execution, common in social science. For example, although the study’s power calculation assumed a baseline rate of 30 events per 1000 admissions, the actual rate proved to be fewer than 7 events per 1000 admissions; thus, the study was ef fect ively underpowered by 500%. Crosscontamination abounded; some control hospitals implemented rapid response protocols, and several study hospi-
[1]
Paul J Sharek,et al.
Effect of a rapid response team on hospital-wide mortality and code rates outside the ICU in a Children's Hospital.
,
2007,
JAMA.
[2]
K. Shojania,et al.
The tension between needing to improve care and knowing how to do it.
,
2007,
The New England journal of medicine.
[3]
F. Davidoff,et al.
What is “quality improvement” and how can it transform healthcare?
,
2007,
Quality and Safety in Health Care.
[4]
Peter J Pronovost,et al.
Rapid response teams--walk, don't run.
,
2006,
JAMA.
[5]
S. Harvey,et al.
Investigating the effectiveness of critical care outreach services: a systematic review
,
2006,
Intensive Care Medicine.
[6]
J. Tibballs,et al.
Reduction of paediatric in-patient cardiac arrest and death with a medical emergency team: preliminary results
,
2005,
Archives of Disease in Childhood.
[7]
F Davidoff,et al.
Toward stronger evidence on quality improvement. Draft publication guidelines: the beginning of a consensus project
,
2005,
Quality and Safety in Health Care.
[8]
K. Hillman,et al.
Introduction of the medical emergency team (MET) system: a cluster-randomised controlled trial
,
2005,
The Lancet.
[9]
B. Bloom.
Crossing the Quality Chasm: A New Health System for the 21st Century
,
2002
.
[10]
L. Kohn,et al.
To Err Is Human : Building a Safer Health System
,
2007
.
[11]
A. Wall,et al.
Book ReviewTo Err is Human: building a safer health system Kohn L T Corrigan J M Donaldson M S Washington DC USA: Institute of Medicine/National Academy Press ISBN 0 309 06837 1 $34.95
,
2000
.
[12]
B. Fisher.
From Halsted to prevention and beyond: advances in the management of breast cancer during the twentieth century.
,
1999,
European journal of cancer.
[13]
T. Brennan,et al.
INCIDENCE OF ADVERSE EVENTS AND NEGLIGENCE IN HOSPITALIZED PATIENTS
,
2008
.
[14]
D M Eddy,et al.
The quality of medical evidence: implications for quality of care.
,
1988,
Health affairs.
[15]
A. Luini,et al.
Comparing Radical Mastectomy with Quadrantectomy, Axillary Dissection, and Radiotherapy in Patients with Small Cancers of the Breast
,
1981
.
[16]
J. Dolovich,et al.
The need for extended treatment of anaphylaxis.
,
1980,
Canadian Medical Association journal.
[17]
R. Brook,et al.
Quality of care assessment: a comparison of five methods of peer review
,
1974
.
[18]
A. Gittelsohn,et al.
Small Area Variations in Health Care Delivery
,
1973,
Science.
[19]
P A LEMBCKE,et al.
Medical auditing by scientific methods; illustrated by major female pelvic surgery.
,
1956,
Journal of the American Medical Association.