Discussion: Why "An estimate of the science-wise false discovery rate and application to the top medical literature" is false.

Jager and Leek have tried to estimate a false-discovery rate (FDR) in abstracts of articles published in five medical journals during 2000-2010. Their approach is flawed in sampling, calculations, and conclusions. It uses a tiny portion of select papers in highly select journals. Randomized controlled trials and systematic reviews (designs with the lowest anticipated false-positive rates) are 52% of the analyzed papers, while these designs account for only 4% in PubMed in the same period. The FDR calculations consider the entire published literature as equivalent to a single genomic experiment where all performed analyses are reported without selection or distortion. However, the data used are the P-values reported in the abstracts of published papers; these P-values are a highly distorted, highly select sample. Besides selective reporting biases, all other biases, in particular confounding in observational studies, are also ignored, while these are often the main drivers for high false-positive rates in the biomedical literature. A reproducibility check of the raw data shows that much of the data Jager and Leek used are either wrong or make no sense: most of the usable data were missed by their script, 94% of the abstracts that reported ≥2 P-values had high correlation/overlap between reported outcomes, and only a minority of P-values corresponded to relevant primary outcomes. The Jager and Leek paper exemplifies the dreadful combination of using automated scripts with wrong methods and unreliable data. Sadly, this combination is common in the medical literature.

[1]  John P A Ioannidis,et al.  Researching Genetic Versus Nongenetic Determinants of Disease: A Comparison and Proposed Unification , 2009, Science Translational Medicine.

[2]  Larry V. Hedges,et al.  Selection Method Approaches , 2006 .

[3]  George Liberopoulos,et al.  Selection in Reported Epidemiological Risks: An Empirical Assessment , 2007, PLoS medicine.

[4]  D. Fanelli “Positive” Results Increase Down the Hierarchy of the Sciences , 2010, PloS one.

[5]  J. Ioannidis Why Most Discovered True Associations Are Inflated , 2008, Epidemiology.

[6]  Douglas G Altman,et al.  Reporting and interpretation of randomized controlled trials with statistically nonsignificant results for primary outcomes. , 2010, JAMA.

[7]  J. Ioannidis,et al.  Magnitude of effects in clinical trials published in high-impact general medical journals. , 2011, International journal of epidemiology.

[8]  C. Begley,et al.  Drug development: Raise standards for preclinical cancer research , 2012, Nature.

[9]  D. Lawlor,et al.  Clustered Environments and Randomized Genes: A Fundamental Distinction between Conventional and Genetic Epidemiology , 2007, PLoS medicine.

[10]  Peter C Gøtzsche,et al.  Believability of relative risks and odds ratios in abstracts: cross sectional study , 2006, BMJ : British Medical Journal.

[11]  J. Ioannidis,et al.  Evaluation of the potential excess of statistically significant findings in published genetic association studies: application to Alzheimer's disease. , 2008, American journal of epidemiology.

[12]  Douglas G Altman,et al.  Epidemiology and reporting of randomised trials published in PubMed journals , 2005, The Lancet.

[13]  Larry V. Hedges,et al.  Estimating Effect Size Under Publication Bias: Small Sample Properties and Robustness of a Random Effects Selection Model , 1996 .

[14]  Jack L. Vevea,et al.  A general linear model for estimating effect size in the presence of publication bias , 1995 .

[15]  J. Ioannidis,et al.  An exploratory test for an excess of significant findings , 2007, Clinical trials.

[16]  F. Prinz,et al.  Believe it or not: how much can we rely on published data on potential drug targets? , 2011, Nature Reviews Drug Discovery.

[17]  J. Ioannidis,et al.  The False-positive to False-negative Ratio in Epidemiologic Studies , 2011, Epidemiology.

[18]  Douglas G Altman,et al.  Discrepancies in sample size calculations and data analyses reported in randomised trials: comparison of publications with protocols , 2008, BMJ : British Medical Journal.

[19]  J. Ioannidis Contradicted and Initially Stronger Effects in Highly Cited Clinical Research , 2005 .

[20]  John P A Ioannidis,et al.  Empirical evaluation of very large treatment effects of medical interventions. , 2012, JAMA.

[21]  J. Brooks Why most published research findings are false: Ioannidis JP, Department of Hygiene and Epidemiology, University of Ioannina School of Medicine, Ioannina, Greece , 2008 .

[22]  Alan F. Karr,et al.  Deming, data and observational studies: A process out of control and needing fixing , 2013 .

[23]  Daniele Fanelli,et al.  Negative results are disappearing from most disciplines and countries , 2011, Scientometrics.

[24]  D. Fanelli Do Pressures to Publish Increase Scientists' Bias? An Empirical Support from US States Data , 2010, PloS one.

[25]  J. Ioannidis,et al.  Quantifying Selective Reporting and the Proteus Phenomenon for Multiple Datasets with Similar Bias , 2011, PloS one.

[26]  S. Stanley Young,et al.  Deming, data and observational studies , 2011 .