BACKGROUND: Multiparametric (mp) magnetic resonance imaging (MRI) has become an important tool for the detection of clinically significant prostate cancer. However, diagnostic accuracy is affected by variability between radiologists. OBJECTIVE: To determine the accuracy and variability in prostate mpMRI interpretation among radiologists, both individually and in teams, in a blinded fashion. DESIGN, SETTING, AND PARTICIPANTS: A study cohort (n=32) was created from our prospective registry of patients who received prostate mpMRI with subsequent biopsy. The cohort was then independently reviewed by four radiologists of varying levels of experience, who assigned a Prostate Imaging Reporting and Data System (PI-RADS) classification, blinded to all clinical information. Consensus interpretation by teams of two radiologists was evaluated after a 12-wk wash-out period. Interpretive accuracy was calculated with various cutoffs for PI-RADS classification and Gleason score. Variability among individual radiologists and teams was calculated using the Fleiss kappa and intraclass correlation coefficient (ICC). RESULTS AND LIMITATIONS: Using PI-RADS 3+/Gleason 7+ (p<0.01) and PI-RADS 4+/Gleason 6+ (p=0.02) as cutoffs, significant differences in accuracy among the four radiologists were noted. At no cutoff for PI-RADS classification or Gleason score did a team read achieve higher accuracy than the most accurate radiologist. The kappa and ICC ranged from 0.22 to 0.29 for the individuals and from 0.16 to 0.21 for the teams (poor agreement). A larger sample size may be needed to adequately power differences in accuracy among individual radiologists. CONCLUSIONS: At various cutoffs for PI-RADS classification and Gleason score, we find significant differences in individual radiologist accuracy, as well as a poor agreement among individual radiologists. Consensus interpretations-as teams of two radiologists-did not improve accuracy or reduce variability. PATIENT SUMMARY: This study investigated radiologist variability and differences in accuracy using multiparametric magnetic resonance imaging for the diagnosis of prostate cancer. Despite attempts to standardize interpretation within the field, we found substantial variability and significant differences in accuracy among individual radiologists.
BACKGROUND: Multiparametric (mp) magnetic resonance imaging (MRI) has become an important tool for the detection of clinically significant prostate cancer. However, diagnostic accuracy is affected by variability between radiologists. OBJECTIVE: To determine the accuracy and variability in prostate mpMRI interpretation among radiologists, both individually and in teams, in a blinded fashion. DESIGN, SETTING, AND PARTICIPANTS: A study cohort (n=32) was created from our prospective registry of patients who received prostate mpMRI with subsequent biopsy. The cohort was then independently reviewed by four radiologists of varying levels of experience, who assigned a Prostate Imaging Reporting and Data System (PI-RADS) classification, blinded to all clinical information. Consensus interpretation by teams of two radiologists was evaluated after a 12-wk wash-out period. Interpretive accuracy was calculated with various cutoffs for PI-RADS classification and Gleason score. Variability among individual radiologists and teams was calculated using the Fleiss kappa and intraclass correlation coefficient (ICC). RESULTS AND LIMITATIONS: Using PI-RADS 3+/Gleason 7+ (p<0.01) and PI-RADS 4+/Gleason 6+ (p=0.02) as cutoffs, significant differences in accuracy among the four radiologists were noted. At no cutoff for PI-RADS classification or Gleason score did a team read achieve higher accuracy than the most accurate radiologist. The kappa and ICC ranged from 0.22 to 0.29 for the individuals and from 0.16 to 0.21 for the teams (poor agreement). A larger sample size may be needed to adequately power differences in accuracy among individual radiologists. CONCLUSIONS: At various cutoffs for PI-RADS classification and Gleason score, we find significant differences in individual radiologist accuracy, as well as a poor agreement among individual radiologists. Consensus interpretations-as teams of two radiologists-did not improve accuracy or reduce variability. PATIENT SUMMARY: This study investigated radiologist variability and differences in accuracy using multiparametric magnetic resonance imaging for the diagnosis of prostate cancer. Despite attempts to standardize interpretation within the field, we found substantial variability and significant differences in accuracy among individual radiologists.
Authors: Rewa R Sood; Wei Shao; Christian Kunder; Nikola C Teslovich; Jeffrey B Wang; Simon J C Soerensen; Nikhil Madhuripan; Anugayathri Jawahar; James D Brooks; Pejman Ghanouni; Richard E Fan; Geoffrey A Sonn; Mirabela Rusu Journal: Med Image Anal Date: 2021-01-23 Impact factor: 8.545
Authors: Armando Stabile; Francesco Giganti; Veeru Kasivisvanathan; Gianluca Giannarini; Caroline M Moore; Anwar R Padhani; Valeria Panebianco; Andrew B Rosenkrantz; Georg Salomon; Baris Turkbey; Geert Villeirs; Jelle O Barentsz Journal: Eur Urol Oncol Date: 2020-03-17
Authors: Jeffrey J Tosoian; Bruce J Trock; Todd M Morgan; Simpa S Salami; Scott A Tomlins; Daniel E Spratt; Javed Siddiqui; Lakshmi P Kunju; Rachel Botbyl; Zoey Chopra; Balaji Pandian; Nicholas W Eyrich; Gary Longton; Yingye Zheng; Ganesh S Palapattu; John T Wei; Yashar S Niknafs; Arul M Chinnaiyan Journal: J Urol Date: 2020-10-20 Impact factor: 7.450
Authors: Lois Kim; Nicholas Boxall; Anne George; Keith Burling; Pete Acher; Jonathan Aning; Stuart McCracken; Toby Page; Vincent J Gnanapragasam Journal: BMC Med Date: 2020-04-17 Impact factor: 8.775