| Literature DB >> 23658692 |
Christopher Baethge1, Jeremy Franklin, Stephan Mertens.
Abstract
BACKGROUND: Peer review is the mainstay of editorial decision making for medical journals. There is a dearth of evaluations of journal peer review with regard to reliability and validity, particularly in the light of the wide variety of medical journals. Studies carried out so far indicate low agreement among reviewers. We present an analysis of the peer review process at a general medical journal, Deutsches Ärzteblatt International. METHODOLOGY/PRINCIPALEntities:
Mesh:
Year: 2013 PMID: 23658692 PMCID: PMC3642182 DOI: 10.1371/journal.pone.0061401
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Reviewer statements.
| Reviewer statement | N (%) |
| Accept without revision | 39 (7.0%) |
| Accept after revision | 408 (73.6%) |
| Reject | 107 (19.3%) |
N = 554 reviews.
Reviewer statements and editorial decisions.
| Reviewer statements by manuscript | % (N) | Editorial decision: acceptance | Relative risk |
| Unanimous: accept or accept after revision | 61.7 (127) | 92.9 (118) | 1.88 [1.55–2.19] |
| Diverging: accept or accept after revision versus reject | 35.0 (72) | 54.2 (39) | 0.62 [0.51–0.76] |
| Unanimous: reject | 3.4 (7) | 0 | 0.0 [0.0–0.56] |
Probability (“risk”) of a manuscript in this group being published compared to the probability of a manuscript in both other groups.
Agreement among reviewers regarding recommendations on the publication of evaluated manuscripts (N = all 529 possible reviewer pairs from 554 reviews on 206 manuscripts).
| Reviewer 2 | |||||
| Accept | Accept after revision | Reject | |||
|
|
| 2 | 36 | 2 | 40 |
|
| 35 | 289 | 62 | 386 | |
|
| 3 | 69 | 31 | 103 | |
| 40 | 394 | 95 | 529 | ||
Cohen's kappa for this 3×3 cross table: 0.059 [95%-CI: −0.016–0.134], Spearman's r: 0.17 (p<0.0001). When “accept” and “accept after revision” were condensed into one variable indicating acceptance Cohen's kappa for the resulting 2×2 cross table was 0.16 [95%-CI: 0.059–0.252], Spearman's r: 0.16 (p = 0.0002).
Designation of reviewers as ‘Reviewer 1’ or ‘Reviewer 2’ was arbitrary.
Figure 1Scatterplot of average reviewer rating (x-axis) and number of citations (y-axis).
N = 157 articles published in Deutsches Ärzteblatt International 2008–2011. X-axis: Reviewer recommendations: 1 denotes accept without revision, 2 denotes accept after minor/major revision, and 3 denotes reject. Spearman's r: −0.06; partial correlation, adjusted for year of publication: −0.03 (n.s.).
Various measurements of reviewer agreement in three studies involving manuscript review at medical journals.
| Study | Journal (# of manuscripts) | Overall agreement | Positive agreement | Negative agreement | Cohen's Kappa | Fleiss' Kappa | Gwet's Kappa |
| Rothwell & Martyn 2000 | Journal A (179) | 57% | 65% | 45% | 0.10 | 0.09 | 0.18 |
| Journal B (116) | 71% | 77% | 60% | 0.37 | 0.37 | 0.46 | |
| Yadollahie et al. 2004 | Iranian J Med Sciences (28) | 46% | 44% | 48% | −0.07 | −0.07 | 0.12 |
| Baethge et al. 2012 | Dtsch Arztebl Int (206) | 75% | 84% | 31% | 0.16 | 0.15 | 0.63 |
Positive agreement denotes agreement regarding acceptance, negative agreement refers to agreement regarding rejection of a manuscript.
Exemplary calculation of the alternative kappa statistic as proposed by Gwet (2002).
| Reviewer 1 | ||||
| Acceptance | Rejection | Total | ||
| Reviewer 2 | Acceptance | A | B | B1 = A+B |
| Rejection | C | D | B2 = C+D | |
| Total | A1 = A+C | A2 = B+D | N | |
Alternative chance agreement probability:
e(y) = 2P1(1-P1).
Approximate chance that a rater (A or B) classifies a subject into category “acceptance”.
P1 = [(A1+B1)/2]/N.
Alternative kappa statistic:
AC1 = [p-e(y)]/[1-e(y)].
Where p = (A+D)/N.