| Literature DB >> 30993432 |
Alejandro Rodriguez-Ruiz1,2, Kristina Lång3, Albert Gubern-Merida2, Jonas Teuwen1, Mireille Broeders4,5, Gisella Gennaro6, Paola Clauser7, Thomas H Helbich7, Margarita Chevalier8, Thomas Mertelmeier9, Matthew G Wallis10, Ingvar Andersson11, Sophia Zackrisson12, Ioannis Sechopoulos1,5, Ritse M Mann13.
Abstract
PURPOSE: To study the feasibility of automatically identifying normal digital mammography (DM) exams with artificial intelligence (AI) to reduce the breast cancer screening reading workload. METHODS AND MATERIALS: A total of 2652 DM exams (653 cancer) and interpretations by 101 radiologists were gathered from nine previously performed multi-reader multi-case receiver operating characteristic (MRMC ROC) studies. An AI system was used to obtain a score between 1 and 10 for each exam, representing the likelihood of cancer present. Using all AI scores between 1 and 9 as possible thresholds, the exams were divided into groups of low- and high likelihood of cancer present. It was assumed that, under the pre-selection scenario, only the high-likelihood group would be read by radiologists, while all low-likelihood exams would be reported as normal. The area under the reader-averaged ROC curve (AUC) was calculated for the original evaluations and for the pre-selection scenarios and compared using a non-inferiority hypothesis.Entities:
Keywords: Artificial intelligence; Breast cancer; Deep learning; Mammography; Screening
Mesh:
Year: 2019 PMID: 30993432 PMCID: PMC6682851 DOI: 10.1007/s00330-019-06186-9
Source DB: PubMed Journal: Eur Radiol ISSN: 0938-7994 Impact factor: 5.315
Fig. 1Distribution of normal (a), cancer (b), and benign exams (c) as a function of AI score, representing the likelihood of cancer present (1–10, 10 means high likelihood of cancer present). The contribution of each dataset to the overall percentage of exams is shown
Fig. 2An example of the nine exams in our study that contained cancer but were assigned an AI score of 1 or 2, the lowest cancer-present likelihood categories. None of the 6 radiologists recalled this exam during the original MRMC study (read without priors), suggesting that the cancer visibility with mammography is poor in these exams (and in fact, the cancer may have been detected by other means)
Fig. 3Proportion (%) of exams that would be excluded from the final sample to be evaluated by the radiologists, using all possible AI scores as thresholds values for pre-selection for reading
Fig. 4ROC curves (a) and change (b) in AUC values of the average of radiologists in the original population, as well as in all possible pre-selected populations (using all possible AI scores as threshold values for pre-selection for reading; if the case is not pre-selected, the radiologist score is converted to the lowest possible cancer suspicion score for the MRMC study). 95% confidence intervals are Bonferroni-corrected