A Powerful Nudge? Presenting Calculable Consequences of Underpowered Research Shifts Incentives Toward Adequately Powered Designs

If psychologists have recognized the pitfalls of underpowered research for decades, why does it persist? Incentives, perhaps: underpowered research benefits researchers individually (increased productivity), but harms science collectively (inflated Type I error rates and effect size estimates but low replication rates). Yet, researchers can selectively reward power at various scientific bottlenecks (e.g., peer review, hiring, funding, and promotion). We designed a stylized thought experiment to evaluate the degree to which researchers consider power and productivity in hiring decisions. Accomplished psychologists chose between a low sample size candidate and a high sample size candidate who were otherwise identical. We manipulated the degree to which participants received information about (1) productivity, (2) sample size, and (3) directly calculable Type I error and replication rates. Participants were intolerant of the negative consequences of low-power research, yet merely indifferent regarding the practices that logically produce those consequences, unless those consequences were made quite explicit.

[1]  F. Al-Shamali,et al.  Author Biographies. , 2015, Journal of social work in disability & rehabilitation.

[2]  Jacob Cohen Statistical Power Analysis , 1992 .

[3]  Lynn Francis Flipping the Self-Control Switch: A Novel Within-Subject Paradigm to Test the Effects of Ego Depletion , 2014 .

[4]  John E. Overall Classical statistical hypotheses testing within the context of Bayesian theory. , 1969 .

[5]  Jacob Cohen,et al.  The statistical power of abnormal-social psychological research: a review. , 1962, Journal of abnormal and social psychology.

[6]  Jeffrey Bowers,et al.  Article Commentary: On the Persistence of Low Power in Psychological Science , 2014, Quarterly journal of experimental psychology.

[7]  M. Bertamini,et al.  Bite-Size Science and Its Undesired Side Effects , 2012, Perspectives on psychological science : a journal of the Association for Psychological Science.

[8]  Brian A. Nosek,et al.  Recommendations for Increasing Replicability in Psychology † , 2013 .

[9]  David Colquhoun,et al.  An investigation of the false discovery rate and the misinterpretation of p-values , 2014, Royal Society Open Science.

[10]  U. Schimmack The ironic effect of significant results on the credibility of multiple-study articles. , 2012, Psychological methods.

[11]  Student BELIEF IN THE LAW OF SMALL NUMBERS , 1994 .

[12]  Jacob Cohen,et al.  A power primer. , 1992, Psychological bulletin.

[13]  Simine Vazire,et al.  The N-Pact Factor: Evaluating the Quality of Empirical Journals with Respect to Sample Size and Statistical Power , 2014, PloS one.

[14]  G. Gigerenzer,et al.  Do studies of statistical power have an effect on the power of studies , 1989 .

[15]  Brian A. Nosek,et al.  Scientific Utopia , 2012, Perspectives on psychological science : a journal of the Association for Psychological Science.

[16]  Brian A. Nosek,et al.  Promoting Transparency in Social Science Research , 2014, Science.

[17]  J. Wicherts,et al.  The Rules of the Game Called Psychological Science , 2012, Perspectives on psychological science : a journal of the Association for Psychological Science.

[18]  Brian A. Nosek,et al.  Power failure: why small sample size undermines the reliability of neuroscience , 2013, Nature Reviews Neuroscience.

[19]  R Core Team,et al.  R: A language and environment for statistical computing. , 2014 .

[20]  Leif D. Nelson,et al.  False-Positive Psychology , 2011, Psychological science.

[21]  Leif D. Nelson,et al.  Life after P-Hacking , 2013 .

[22]  K. Vohs,et al.  Psychology as the Science of Self-reports and Finger Movements Whatever Happened to Actual Behavior? , 2022 .

[23]  C. F. Bond,et al.  One Hundred Years of Social Psychology Quantitatively Described , 2003 .