Reviewer agreement trends from four years of electronic submissions of conference abstract

BackgroundThe purpose of this study was to determine the inter-rater agreement between reviewers on the quality of abstract submissions to an annual national scientific meeting (Canadian Association of Emergency Physicians; CAEP) to identify factors associated with low agreement.MethodsAll abstracts were submitted using an on-line system and assessed by three volunteer CAEP reviewers blinded to the abstracts' source. Reviewers used an on-line form specific for each type of study design to score abstracts based on nine criteria, each contributing from two to six points toward the total (maximum 24). The final score was determined to be the mean of the three reviewers' scores using Intraclass Correlation Coefficient (ICC).Results495 Abstracts were received electronically during the four-year period, 2001 – 2004, increasing from 94 abstracts in 2001 to 165 in 2004. The mean score for submitted abstracts over the four years was 14.4 (95% CI: 14.1–14.6). While there was no significant difference between mean total scores over the four years (p = 0.23), the ICC increased from fair (0.36; 95% CI: 0.24–0.49) to moderate (0.59; 95% CI: 0.50–0.68). Reviewers agreed less on individual criteria than on the total score in general, and less on subjective than objective criteria.ConclusionThe correlation between reviewers' total scores suggests general recognition of "high quality" and "low quality" abstracts. Criteria based on the presence/absence of objective methodological parameters (i.e., blinding in a controlled clinical trial) resulted in higher inter-rater agreement than the more subjective and opinion-based criteria. In future abstract competitions, defining criteria more objectively so that reviewers can base their responses on empirical evidence may lead to increased consistency of scoring and, presumably, increased fairness to submitters.

[1]  W. McIlroy,et al.  Should unpublished data be included in meta-analyses? Current convictions and controversies. , 1993, JAMA.

[2]  Cindy Farquhar,et al.  3 The Cochrane Library , 1996 .

[3]  K. Dickersin,et al.  Full publication of results initially presented in abstracts. A meta-analysis. , 1994 .

[4]  J. Fleiss,et al.  Intraclass correlations: uses in assessing rater reliability. , 1979, Psychological bulletin.

[5]  P. Rothwell,et al.  Reproducibility of peer review in clinical neuroscience. Is agreement between reviewers any greater than would be expected by chance alone? , 2000, Brain : a journal of neurology.

[6]  P. Tugwell,et al.  Does the inclusion of grey literature influence estimates of intervention effectiveness reported in meta-analyses? , 2000, The Lancet.

[7]  P. Vargha,et al.  A critical discussion of intraclass correlation coefficients. , 1997, Statistics in medicine.

[8]  Kevin R. Murphy,et al.  Distributional ratings, judgment decomposition, and their impact on interrater agreement and rating accuracy. , 1990 .

[9]  E. von Elm,et al.  Full publication of results initially presented in abstracts. , 2007, The Cochrane database of systematic reviews.

[10]  Domenic V. Cicchetti,et al.  A Statistical Analysis of Reviewer Agreement and Bias in Evaluating Medical Abstracts 1 , 1976, The Yale journal of biology and medicine.

[11]  A. Wu,et al.  How reliable is peer review of scientific abstracts? , 1993, Journal of general internal medicine.

[12]  Antje Timmer,et al.  BMC Medical Research Methodology Research article Design of Phase II cancer trials evaluating survival , 2003 .

[13]  Tom Fahey,et al.  Inter-rater agreement in the scoring of abstracts submitted to a primary care research conference , 2002, BMC health services research.

[14]  D. Cicchetti,et al.  Improving participation and interrater agreement in scoring Ambulatory Pediatric Association abstracts. How well have we succeeded? , 1996, Archives of pediatrics & adolescent medicine.

[15]  Mph Albert W. Wu MD,et al.  How reliable is peer review of scientific abstracts? , 2007, Journal of General Internal Medicine.