An illustration of reproducibility in neuroscience research in the absence of selective reporting

The problem of poor reproducibility of scientific findings has received much attention over recent years, in a variety of fields including psychology and neuroscience. The problem has been partly attributed to publication bias and unwanted practices such as p-hacking. Low statistical power in individual studies is also understood to be an important factor. In a recent multi-site collaborative study, we mapped brain anatomical left-right asymmetries for regional measures of surface area and cortical thickness, in 99 MRI datasets from around the world, for a total of over 17,000 participants. In the present study, we re-visited these hemispheric effects from the perspective of reproducibility. Within each dataset, we considered that an effect had been reproduced when it matched the meta-analytic effect from the 98 other datasets, in terms of effect direction and uncorrected significance at p<0.05. In this sense, the results within each dataset were viewed as coming from separate studies in an ‘ideal publishing environment’, i.e. free from selective reporting and p hacking. We found an average reproducibility rate per dataset, over all effects, of 63.2% (SD = 22.9%, min = 22.2%, max = 97.0%). As expected, reproducibility was higher for larger effects and in larger datasets. There is clearly substantial room to improve reproducibility in brain MRI research through increasing statistical power. These findings constitute an empirical illustration of reproducibility in the absence of publication bias or p hacking, when assessing realistic biological effects in heterogeneous neuroscience data, and given typically-used sample sizes.

[1]  Michael C. Frank,et al.  Estimating the reproducibility of psychological science , 2015, Science.

[2]  Thomas E. Nichols,et al.  Scanning the horizon: towards transparent and reproducible neuroimaging research , 2016, Nature Reviews Neuroscience.

[3]  Brian A. Nosek,et al.  Publication and other reporting biases in cognitive sciences: detection, prevalence, and prevention , 2014, Trends in Cognitive Sciences.

[4]  G. Loewenstein,et al.  Measuring the Prevalence of Questionable Research Practices With Incentives for Truth Telling , 2012, Psychological science.

[5]  Bruce Fischl,et al.  FreeSurfer , 2012, NeuroImage.

[6]  M. Baker 1,500 scientists lift the lid on reproducibility , 2016, Nature.

[7]  Leif D. Nelson,et al.  False-Positive Psychology , 2011, Psychological science.

[8]  J. Ioannidis Why Most Discovered True Associations Are Inflated , 2008, Epidemiology.

[9]  Reginald B. Adams,et al.  Investigating Variation in Replicability: A “Many Labs” Replication Project , 2014 .

[10]  J. Ioannidis Why Most Published Research Findings Are False , 2005, PLoS medicine.

[11]  Thomas E. Nichols,et al.  Exploring fMRI Results Space: 31 Variants of an fMRI Analysis in AFNI, FSL, and SPM , 2016, Front. Neuroinform..

[12]  M. Sheridan,et al.  Sample composition alters associations between age and brain structure , 2017, Nature Communications.

[13]  J. Wicherts,et al.  The Rules of the Game Called Psychological Science , 2012, Perspectives on psychological science : a journal of the Association for Psychological Science.

[14]  Brian A. Nosek,et al.  Power failure: why small sample size undermines the reliability of neuroscience , 2013, Nature Reviews Neuroscience.

[15]  Gary H Glover,et al.  Estimating sample size in functional MRI (fMRI) neuroimaging studies: Statistical power analyses , 2002, Journal of Neuroscience Methods.

[16]  S. Eickhoff,et al.  Empirical examination of the replicability of associations between brain structure and psychological variables , 2018, bioRxiv.

[17]  Luca Turella,et al.  Variability in the analysis of a single neuroimaging dataset by many teams , 2019, Nature.

[18]  Michael C. Frank,et al.  A practical guide for transparency in psychological science , 2018 .

[19]  Brian A. Nosek,et al.  Promoting an open research culture , 2015, Science.

[20]  Christopher D. Chambers,et al.  Redefine statistical significance , 2017, Nature Human Behaviour.

[21]  J. Ioannidis Why Most Published Research Findings Are False , 2019, CHANCE.

[22]  Thomas E. Nichols,et al.  The ENIGMA Consortium: large-scale collaborative analyses of neuroimaging and genetic data , 2014, Brain Imaging and Behavior.

[23]  Sander Greenland,et al.  Remove, rather than redefine, statistical significance , 2017, Nature Human Behaviour.

[24]  F. Prinz,et al.  Believe it or not: how much can we rely on published data on potential drug targets? , 2011, Nature Reviews Drug Discovery.

[25]  Karolinska Schizophrenia Mapping cortical brain asymmetry in 17,141 healthy individuals worldwide via the ENIGMA consortium , 2018 .

[26]  Reginald B. Adams,et al.  Many Labs 2: Investigating Variation in Replicability Across Sample and Setting , 2018 .