Craig K Abbey1, Brandon D Gallas2, John M Boone3, Loren T Niklason4, Lubomir M Hadjiiski5, Berkman Sahiner2, Frank W Samuelson2. 1. Department of Psychological and Brain Sciences, University of California, Santa Barbara, CA 93106. Electronic address: abbey@psych.ucsb.edu. 2. US Food and Drug Administration, Center for Devices and Radiological Health, Office of Science and Engineering Laboratories, Silver Spring, MD. 3. Department of Radiology, UC Davis Medical Center, Sacramento, CA. 4. Hologic Inc., Bedford, MA. 5. Department of Radiology, University of Michigan Comprehensive Cancer Center, Ann Arbor, MI.
Abstract
RATIONALE AND OBJECTIVES: Our objective is to determine whether expected utility (EU) and the area under the receiver operator characteristic (AUC) are consistent with one another as endpoints of observer performance studies in mammography. These two measures characterize receiver operator characteristic performance somewhat differently. We compare these two study endpoints at the level of individual reader effects, statistical inference, and components of variance across readers and cases. MATERIALS AND METHODS: We reanalyze three previously published laboratory observer performance studies that investigate various x-ray breast imaging modalities using EU and AUC. The EU measure is based on recent estimates of relative utility for screening mammography. RESULTS: The AUC and EU measures are correlated across readers for individual modalities (r = 0.93) and differences in modalities (r = 0.94 to 0.98). Statistical inference for modality effects based on multi-reader multi-case analysis is very similar, with significant results (P < .05) in exactly the same conditions. Power analyses show mixed results across studies, with a small increase in power on average for EU that corresponds to approximately a 7% reduction in the number of readers. Despite a large number of crossing receiver operator characteristic curves (59% of readers), modality effects only rarely have opposite signs for EU and AUC (6%). CONCLUSIONS: We do not find any evidence of systematic differences between EU and AUC in screening mammography observer studies. Thus, when utility approaches are viable (i.e., an appropriate value of relative utility exists), practical effects such as statistical efficiency may be used to choose study endpoints.
RATIONALE AND OBJECTIVES: Our objective is to determine whether expected utility (EU) and the area under the receiver operator characteristic (AUC) are consistent with one another as endpoints of observer performance studies in mammography. These two measures characterize receiver operator characteristic performance somewhat differently. We compare these two study endpoints at the level of individual reader effects, statistical inference, and components of variance across readers and cases. MATERIALS AND METHODS: We reanalyze three previously published laboratory observer performance studies that investigate various x-ray breast imaging modalities using EU and AUC. The EU measure is based on recent estimates of relative utility for screening mammography. RESULTS: The AUC and EU measures are correlated across readers for individual modalities (r = 0.93) and differences in modalities (r = 0.94 to 0.98). Statistical inference for modality effects based on multi-reader multi-case analysis is very similar, with significant results (P < .05) in exactly the same conditions. Power analyses show mixed results across studies, with a small increase in power on average for EU that corresponds to approximately a 7% reduction in the number of readers. Despite a large number of crossing receiver operator characteristic curves (59% of readers), modality effects only rarely have opposite signs for EU and AUC (6%). CONCLUSIONS: We do not find any evidence of systematic differences between EU and AUC in screening mammography observer studies. Thus, when utility approaches are viable (i.e., an appropriate value of relative utility exists), practical effects such as statistical efficiency may be used to choose study endpoints.
Authors: Brandon D Gallas; Heang-Ping Chan; Carl J D'Orsi; Lori E Dodd; Maryellen L Giger; David Gur; Elizabeth A Krupinski; Charles E Metz; Kyle J Myers; Nancy A Obuchowski; Berkman Sahiner; Alicia Y Toledano; Margarita L Zuley Journal: Acad Radiol Date: 2012-02-03 Impact factor: 3.173
Authors: Robert M Nishikawa; Suddhasatta Acharyya; Constantine Gatsonis; Etta D Pisano; Elodia B Cole; Helga S Marques; Carl J D'Orsi; Dione M Farria; Kalpana M Kanal; Mary C Mahoney; Murray Rebner; Melinda J Staiger Journal: Radiology Date: 2009-04 Impact factor: 11.105
Authors: Elizabeth A Rafferty; Jeong Mi Park; Liane E Philpotts; Steven P Poplack; Jules H Sumkin; Elkan F Halpern; Loren T Niklason Journal: Radiology Date: 2012-11-20 Impact factor: 11.105
Authors: Craig K Abbey; Yirong Wu; Elizabeth S Burnside; Adam Wunderlich; Frank W Samuelson; John M Boone Journal: Proc SPIE Int Soc Opt Eng Date: 2016-03-24