Reliability of Editors' Subjective Quality Ratings of Peer Reviews of Manuscripts
Overview
Authors
Affiliations
Context: Quality of reviewers is crucial to journal quality, but there are usually too many for editors to know them all personally. A reliable method of rating them (for education and monitoring) is needed.
Objective: Whether editors' quality ratings of peer reviewers are reliable and how they compare with other performance measures.
Design: A 3.5-year prospective observational study.
Setting: Peer-reviewed journal.
Participants: All editors and peer reviewers who reviewed at least 3 manuscripts.
Main Outcome Measures: Reviewer quality ratings, individual reviewer rate of recommendation for acceptance, congruence between reviewer recommendation and editorial decision (decision congruence), and accuracy in reporting flaws in a masked test manuscript.
Interventions: Editors rated the quality of each review on a subjective 1 to 5 scale.
Results: A total of 4161 reviews of 973 manuscripts by 395 reviewers were studied. The within-reviewer intraclass correlation was 0.44 (P<.001), indicating that 20% of the variance seen in the review ratings was attributable to the reviewer. Intraclass correlations for editor and manuscript were only 0.24 and 0.12, respectively. Reviewer average quality ratings correlated poorly with the rate of recommendation for acceptance (R=-0.34) and congruence with editorial decision (R=0.26). Among 124 reviewers of the fictitious manuscript, the mean quality rating for each reviewer was modestly correlated with the number of flaws they reported (R=0.53). Highly rated reviewers reported twice as many flaws as poorly rated reviewers.
Conclusions: Subjective editor ratings of individual reviewers were moderately reliable and correlated with reviewer ability to report manuscript flaws. Individual reviewer rate of recommendation for acceptance and decision congruence might be thought to be markers of a discriminating (ie, high-quality) reviewer, but these variables were poorly correlated with editors' ratings of review quality or the reviewer's ability to detect flaws in a fictitious manuscript. Therefore, they cannot be substituted for actual quality ratings by editors.
Love J, Messman A, Ilgen J, Merritt C, Coates W, Ander D West J Emerg Med. 2024; 25(2):254-263.
PMID: 38596927 PMC: 11000557. DOI: 10.5811/westjem.18432.
A structured, journal-led peer-review mentoring program enhances peer review training.
Lyons-Warren A, Aamodt W, Pieper K, Strowd R Res Integr Peer Rev. 2024; 9(1):3.
PMID: 38454514 PMC: 10921741. DOI: 10.1186/s41073-024-00143-x.
Reviewer training for improving grant and journal peer review.
Hesselberg J, Dalsbo T, Stromme H, Svege I, Fretheim A Cochrane Database Syst Rev. 2023; 11:MR000056.
PMID: 38014743 PMC: 10683016. DOI: 10.1002/14651858.MR000056.pub2.
Problems with Peer Review Shine a Light on Gaps in Scientific Training.
Proctor D, Dada N, Serquina A, Willett J mBio. 2023; 14(3):e0318322.
PMID: 37052475 PMC: 10294629. DOI: 10.1128/mbio.03183-22.
Peer review analyze: A novel benchmark resource for computational analysis of peer reviews.
Ghosal T, Kumar S, Bharti P, Ekbal A PLoS One. 2022; 17(1):e0259238.
PMID: 35085252 PMC: 8794172. DOI: 10.1371/journal.pone.0259238.