Issues relating to study design and risk of bias when including non‐randomized studies in systematic reviews on the effects of interventions

Non‐randomized studies may provide valuable evidence on the effects of interventions. They are the main source of evidence on the intended effects of some types of interventions and often provide the only evidence about the effects of interventions on long‐term outcomes, rare events or adverse effects. Therefore, systematic reviews on the effects of interventions may include various types of non‐randomized studies. In this second paper in a series, we address how review authors might articulate the particular non‐randomized study designs they will include and how they might evaluate, in general terms, the extent to which a particular non‐randomized study is at risk of important biases. We offer guidance for describing and classifying different non‐randomized designs based on specific features of the studies in place of using non‐informative study design labels. We also suggest criteria to consider when deciding whether to include non‐randomized studies. We conclude that a taxonomy of study designs based on study design features is needed. Review authors need new tools specifically to assess the risk of bias for some non‐randomized designs that involve a different inferential logic compared with parallel group trials. Copyright © 2012 John Wiley & Sons, Ltd.

[1]  Oxford Centre for Evidence-based Medicine Levels of Evidence (January 2001) , 2014 .

[2]  P. Tugwell,et al.  The Newcastle-Ottawa Scale (NOS) for Assessing the Quality of Nonrandomised Studies in Meta-Analyses , 2014 .

[3]  Laurie M. Anderson,et al.  Issues relating to selective reporting when including non‐randomized studies in systematic reviews on the effects of healthcare interventions , 2013, Research synthesis methods.

[4]  P. Tugwell,et al.  An introduction to methodological issues when including non‐randomised studies in systematic reviews on the effects of interventions , 2013, Research synthesis methods.

[5]  Simon G Thompson,et al.  Issues relating to confounding and meta‐analysis when including non‐randomized studies in systematic reviews on the effects of interventions , 2013, Research synthesis methods.

[6]  Ross J. Harris,et al.  Influence of reported study design characteristics on intervention effect estimates from randomised controlled trials: combined analysis of meta-epidemiological studies. , 2012, Health technology assessment.

[7]  P. Santaguida,et al.  Testing a tool for the classification of study designs in systematic reviews of interventions and exposures showed moderate reliability and low accuracy. , 2011, Journal of clinical epidemiology.

[8]  Su Golder,et al.  Meta-analyses of Adverse Effects Data Derived from Randomised Controlled Trials as Compared to Observational Studies: Methodological Overview , 2011, PLoS medicine.

[9]  Josiane Bonnefoy,et al.  Assessing equity in systematic reviews: realising the recommendations of the Commission on Social Determinants of Health , 2010, BMJ : British Medical Journal.

[10]  Michele Tarsilla Cochrane Handbook for Systematic Reviews of Interventions , 2010, Journal of MultiDisciplinary Evaluation.

[11]  D. Moher,et al.  Randomized trials published in some Chinese journals: how many are randomized? , 2009, Trials.

[12]  Douglas G. Altman,et al.  Models for potentially biased evidence in meta‐analysis using empirically based priors , 2009 .

[13]  Joshua D. Angrist,et al.  Mostly Harmless Econometrics: An Empiricist's Companion , 2008 .

[14]  Bruce M Psaty,et al.  Benefits and risks of drug treatments: how to combine the best evidence on benefits with the best data about adverse effects. , 2008, JAMA.

[15]  Jonathan J Deeks,et al.  Including Non‐Randomized Studies , 2008 .

[16]  Gordon H. Guyatt,et al.  Presenting Results and ‘Summary of Findings’ Tables , 2008 .

[17]  D. Altman,et al.  Assessing Risk of Bias in Included Studies , 2008 .

[18]  William H. Yeaton,et al.  Shaping Policies Related to Developmental Education: An Evaluation Using the Regression-Discontinuity Design , 2006 .

[19]  V Demicheli,et al.  Efficacy and effectiveness of influenza vaccines in elderly people: a systematic review , 2005, The Lancet.

[20]  N. Low,et al.  HIV and male circumcision--a systematic review with assessment of the quality of studies. , 2005, The Lancet. Infectious diseases.

[21]  V Demicheli,et al.  Assessment of the efficacy and effectiveness of influenza vaccines in healthy children: systematic review , 2005, The Lancet.

[22]  F. Song,et al.  Evaluating non-randomised intervention studies. , 2003, Health technology assessment.

[23]  Julie Kreyenbuhl,et al.  Assessment of independent effect of olanzapine and risperidone on risk of diabetes among patients with schizophrenia: population based nested case-control study , 2002, BMJ : British Medical Journal.

[24]  J. Grimshaw,et al.  Getting evidence into practice: the work of the Cochrane Effective Practice and Organization of care Group (EPOC). , 2001, The Journal of continuing education in the health professions.

[25]  A. Moxey,et al.  Agreement between randomized and non-randomized studies: the effects of bias and confounding , 2001 .

[26]  W. Shadish,et al.  Experimental and Quasi-Experimental Designs for Generalized Causal Inference , 2001 .

[27]  B. Psaty,et al.  Assessment and Control for Confounding by Indication in Observational Studies , 1999, Journal of the American Geriatrics Society.

[28]  J. Sterne,et al.  Methods for evaluating area-wide and organisation-based interventions in health and health care: a systematic review. , 1999, Health technology assessment.

[29]  R. D'Agostino Adjustment Methods: Propensity Score Methods for Bias Reduction in the Comparison of a Treatment to a Non‐Randomized Control Group , 2005 .

[30]  N. Black,et al.  The feasibility of creating a checklist for the assessment of the methodological quality both of randomised and non-randomised studies of health care interventions. , 1998, Journal of epidemiology and community health.

[31]  J. Grimshaw,et al.  North of England evidence based guidelines development project: methods of guideline development , 1996, BMJ.

[32]  Joshua D. Angrist,et al.  Identification of Causal Effects Using Instrumental Variables , 1993 .

[33]  Stephen B. Gruber,et al.  Clinical Epidemiology: The Architecture of Clinical Research , 1986, The Yale Journal of Biology and Medicine.

[34]  A. B. Hill The Environment and Disease: Association or Causation? , 1965, Proceedings of the Royal Society of Medicine.

[35]  D. Campbell,et al.  Regression-Discontinuity Analysis: An Alternative to the Ex-Post Facto Experiment , 1960 .