The Science of Using Science: Towards an Understanding of the Threats to Scaling Experiments

Policymakers are increasingly turning to insights gained from the experimental method as a means of informing public policies. Whether—and to what extent—insights from a research study scale to the level of the broader public is, in many situations, based on blind faith. This scale-up problem can lead to a vast waste of resources, a missed opportunity to improve people’s lives, and a diminution in the public’s trust in the scientific method’s ability to contribute to policymaking. This study provides a theoretical lens to deepen our understanding of the science of how to use science. Through a simple model, we highlight three elements of the scale-up problem: (1) when does evidence become actionable (appropriate statistical inference); (2) properties of the population; and (3) properties of the situation. We argue that until these three areas are fully understood and recognized by researchers and policymakers, the threats to scalability will render any scaling exercise as particularly vulnerable. In this way, our work represents a challenge to empiricists to estimate the nature and extent of how important the various threats to scalability are in practice, and to implement those in their original research.

[1]  Elaine Walsh,et al.  Factors affecting patient participation in clinical trials in Ireland: A narrative review , 2016, Contemporary clinical trials communications.

[2]  Petra Kaufmann,et al.  Experimental And Quasi Experimental Designs For Research , 2016 .

[3]  P. V. von Hippel,et al.  Does a Successful Randomized Experiment Lead to Successful Policy? Project Challenge and What Happened in Tennessee After Project STAR , 2018 .

[4]  Karthik Muralidharan,et al.  Experimentation at Scale , 2017 .

[5]  Yves Zenou,et al.  Are Estimates of Early Education Programs Too Pessimistic? Evidence from a Large-Scale Field Experiment that Causally Measures Neighbor Effects , 2019, SSRN Electronic Journal.

[6]  J. Brooks Why most published research findings are false: Ioannidis JP, Department of Hygiene and Epidemiology, University of Ioannina School of Medicine, Ioannina, Greece , 2008 .

[7]  John A. List,et al.  What Can We Learn from Experiments? Understanding the Threats to the Scalability of Experimental Results , 2017 .

[8]  John A. List,et al.  The Market for Charitable Giving , 2011 .

[9]  A. Kilbourne,et al.  Implementing evidence-based interventions in health care: application of the replicating effective programs framework , 2007, Implementation science : IS.

[10]  James J. Heckman,et al.  Characterizing Selection Bias Using Experimental Data , 1998 .

[11]  Lauren H. Supplee,et al.  Opportunities and Challenges in Evidence-based Social Policy , 2015 .

[12]  Diane Paulsell,et al.  Supporting Quality in HomeBased Child Care Initiative Design and Evaluation Options , 2010 .

[13]  A. Biglan,et al.  Research design issues for evaluating complex multicomponent interventions in neighborhoods and communities , 2016, Translational behavioral medicine.

[14]  David A. Chambers,et al.  The dynamic sustainability framework: addressing the paradox of sustainment amid ongoing change , 2013, Implementation Science.

[15]  D. Hunninghake,et al.  Recruitment for controlled clinical trials: literature summary and annotated bibliography. , 1997, Controlled clinical trials.

[16]  Thomas D. Cook,et al.  Standards of Evidence for Efficacy, Effectiveness, and Scale-up Research in Prevention Science: Next Generation , 2015, Prevention Science.

[17]  J. Hektner,et al.  Can Evidence-Based Prevention Programs be Sustained in Community Practice Settings? The Early Risers’ Advanced-Stage Effectiveness Trial , 2006, Prevention Science.

[18]  Omar Al-Ubaydli,et al.  Scaling for Economists: Lessons from the Non-Adherence Problem in the Medical Literature. , 2017, The journal of economic perspectives : a journal of the American Economic Association.

[19]  O. Bandiera,et al.  Losing Prosociality in the Quest for Talent? Sorting, Selection, and Productivity in the Delivery of Public Services , 2018, American Economic Review.

[20]  C. Achilles The Lasting Benefits Study (LBS) in Grades 4 and 5 (1990-1991): A Legacy from Tennessee's Four-Year (K-3) Class-Size Study (1985-1989), Project STAR. Paper #7. , 1993 .

[21]  Jon Baron,et al.  A Brief History of Evidence-Based Policy , 2018, The ANNALS of the American Academy of Political and Social Science.

[22]  Steven D. Levitt,et al.  Field Experiments in Economics: the Past, the Present, and the Future , 2008 .

[23]  J. Ioannidis,et al.  Why Current Publication Practices May Distort Science , 2008, PLoS medicine.

[24]  Lori A. Roggman,et al.  Who Drops Out of Early Head Start Home Visiting Programs? , 2008 .

[25]  Michael J. Weiss,et al.  How Much Do the Effects of Education and Training Programs Vary Across Sites? Evidence From Past Multisite Randomized Trials , 2017 .

[26]  John A. List,et al.  Testing the Theory of Multitasking: Evidence from a Natural Field Experiment in Chinese Factories , 2013, SSRN Electronic Journal.

[27]  Guideline for the study and evaluation of gender differences in the clinical evaluation of drugs; notice. , 1993, Federal register.

[28]  Hayden B Bosworth,et al.  Ingredients of successful interventions to improve medication adherence. , 2013, JAMA.

[29]  Howard S. Bloom,et al.  A Conceptual Framework for Studying the Sources of Variation in Program Effects , 2013 .

[30]  Harini Kannan,et al.  From Proof of Concept to Scalable Policies: Challenges and Solutions, with an Application , 2016 .

[31]  Jens Ludwig,et al.  The Economics of Scale-Up , 2017 .

[32]  Christopher Jepsen,et al.  Class Size Reduction and Student Achievement: The Potential Tradeoff between Teacher Quality and Class Size , 2012 .

[33]  Aleta L. Meyer,et al.  The Intersection Between Prevention Science and Evidence-Based Policy: How the SPR Evidence Standards Support Human Services Prevention Programs , 2015, Prevention Science.

[34]  N. Leighl,et al.  Do the American Society of Clinical Oncology Value Framework and the European Society of Medical Oncology Magnitude of Clinical Benefit Scale Measure the Same Construct of Clinical Benefit? , 2017, Journal of clinical oncology : official journal of the American Society of Clinical Oncology.

[35]  Timothy J. Lewis,et al.  Scaling Up School-Wide Positive Behavioral Interventions and Supports , 2014 .

[36]  Catherine S Tamis-LeMonda,et al.  Mother-child bookreading in low-income families: correlates and outcomes during the first three years of life. , 2006, Child development.

[37]  Kathleen L. McGinn,et al.  Negotiating a Better Future: How Interpersonal Skills Facilitate Inter-Generational Investment , 2018 .

[38]  R. Gersten,et al.  The Impact of Collaborative Strategic Reading on the Reading Comprehension of Grade 5 Students in Linguistically Diverse Schools. Final Report. NCEE 2011-4001. , 2011 .

[39]  A. Deaton,et al.  Understanding and Misunderstanding Randomized Controlled Trials , 2016, Social science & medicine.

[40]  John A. List,et al.  One swallow doesn’t make a summer: new evidence on anchoring effects , 2014 .

[41]  Daniel Hind,et al.  A rapid review indicated higher recruitment rates in treatment trials than in prevention trials. , 2015, Journal of clinical epidemiology.

[42]  John A. List,et al.  Using Lotteries to Finance Public Goods: Theory and Experimental Evidence , 2007 .

[43]  John A. List,et al.  Do Natural Field Experiments Afford Researchers More or Less Control than Laboratory Experiments? a Simple Model , 2015 .

[44]  Howard S. Bloom,et al.  Learning About and From a Distribution of Program Impacts Using Multisite Trials , 2015 .