We Have Met the Enemy and He Is Us

Numerous authors have demonstrated that when multiple institutional review boards (IRBs) review the same protocol, the results are quite variable (Dilts et al. 2009; Dziak et al. 2005; Green et al. 2006; Greene et al. 2006; Helfand et al. 2009; Kimberly et al. 2006; Mansbach et al. 2007; McWilliams et al. 2003; Silverman et al. 2001; Stair et al. 2001; Stark et al. 2010). However, the federal regulations governing the conduct of human subjects research are gray, by either design or accident, and it is therefore to be expected that reasonable people with differing interpretations will reach disparate conclusions about whether or not a given study meets the criteria for approval (Sayers 2007). Klitzman (2011) provides anecdotal evidence that lends credence to another aspect of IRB variability long known to clinical investigators, namely, that there are differences between the IRBs within institutions. The Klitzman paper is fundamentally based on anecdotes. There is no reason why clinical research into IRB decision making shouldn’t be held to the same standards as those for other types of clinical research. As Halpern (2005) advocated recently, there is a need for evidence-based bioethics to inform decisions. From an epidemiologist’s perspective, large randomized clinical trials rank highest in strength of evidence, followed by smaller clinical trials, cohort studies, case-control studies, and then anecdotal evidence at the bottom. Depending on one’s perspective, meta-analyses combining the results of several trials would be at or near the top of the hierarchy of evidence. Even when a study has a high degree of internal validity, the results from a study may not be generalizable to one’s own situation, patient population, and institution. Given the low position within the hierarchy of evidence occupied by this current report, the results and conclusions reported by Klitzman should be viewed with caution and skepticism. If cause and effect are beyond the realm of the anecdotal study design, then why bother with this type of research? Not all research needs to involve confirmatory designs with formal hypothesis testing. Other exploratory designs are valuable when aimed at generating hypotheses and directions for future research. The power of this paper, then, is not in the amalgamation of the results into a

[1]  R. Platt Multicenter Airway Research Collaboration , 2000 .

[2]  S. Hull,et al.  Variability among institutional review boards’ decisions within the context of a multicenter trial , 2001, Critical care medicine.

[3]  M. Radeos,et al.  Variation in institutional review board responses to a standard protocol for a multicenter clinical trial. , 2001, Academic emergency medicine : official journal of the Society for Academic Emergency Medicine.

[4]  Ada Hamosh,et al.  Problematic variation in local institutional review of a multicenter genetic epidemiology study. , 2003, JAMA.

[5]  Seema K. Shah,et al.  How do institutional review boards apply the federal risk and benefit standards for pediatric research? , 2004, JAMA.

[6]  Roger T. Anderson,et al.  Variations among Institutional Review Board reviews in a multisite health services research study. , 2005, Health services research.

[7]  E. Emanuel,et al.  Quantifying the federal minimal risk standard: implications for pediatric research without a prospect of direct benefit. , 2005, JAMA.

[8]  S. Halpern Towards evidence based bioethics , 2005, BMJ : British Medical Journal.

[9]  J. Lowery,et al.  Impact of institutional review board practice variation on observational health services research. , 2006, Health services research.

[10]  C. Feudtner,et al.  Variation in Standards of Research Compensation and Child Assent Practices: A Comparison of 69 Institutional Review Board–Approved Informed Permission and Assent Forms for 3 Multicenter Pediatric Clinical Trials , 2006, Pediatrics.

[11]  Sarah M. Greene,et al.  A review finds that multicenter studies face substantial challenges but strategies exist to achieve Institutional Review Board approval. , 2006, Journal of clinical epidemiology.

[12]  G. Sayers Should research ethics committees be told how to think? , 2007, Journal of Medical Ethics.

[13]  J. Lantos Should institutional review board decisions be evidence-based? , 2007, Archives of pediatrics & adolescent medicine.

[14]  C. Camargo,et al.  Variation in institutional review board responses to a standard, observational, pediatric research protocol. , 2007, Academic emergency medicine : official journal of the Society for Academic Emergency Medicine.

[15]  J. Thigpen Steps and Time to Process Clinical Trials at the Cancer Therapy Evaluation Program , 2009 .

[16]  C. Roehrborn,et al.  Variation in institutional review board responses to a standard protocol for a multicenter randomized, controlled surgical trial. , 2009, The Journal of urology.

[17]  J. Tyson,et al.  Variation among institutional review boards in evaluating the design of a multicenter randomized trial , 2010, Journal of Perinatology.

[18]  A. Tait Weighing the risks and benefits of pediatric research: a risky business. , 2010, Archives of pediatrics & adolescent medicine.

[19]  J. Menikoff,et al.  The paradoxical problem with multiple-IRB review. , 2010, The New England journal of medicine.

[20]  E. Emanuel,et al.  Evaluating the risks of clinical research. , 2010, JAMA.

[21]  R. Klitzman The Myth of Community Differences as the Cause of Variations Among IRBs , 2011, AJOB primary research.