Literature DB >> 35111337

Breast cancer detection across dense and non-dense breasts: Markers of diagnostic confidence and efficacy.

Ibrahim Hadadi1,2, William Rae1, Jillian Clarke1, Mark McEntee1,3, Ernest Ekpo1,4.   

Abstract

BACKGROUND: The impact of radiologists' characteristics has become a major focus of recent research. However, the markers of diagnostic efficacy and confidence in dense and non-dense breasts are poorly understood.
PURPOSE: This study aims to assess the relationship between radiologists' characteristics and diagnostic performance across dense and non-dense breasts.
MATERIALS AND METHODS: Radiologists specialising in breast imaging (n = 128) who had 0.5-40 (13±10.6) years of experience reading mammograms were recruited. Participants independently interpreted a test set containing 60 digital mammograms (40 normal and 20 abnormal) with similarly distributed breast densities. Diagnostic performance measures were analysed via Jamovi software (version 1.6.22).
RESULTS: In dense breasts, breast-imaging fellowship completion significantly improved specificity (p = 0.004), location sensitivity (p = 0.01) and the area under the curve (AUC) of the receiver operating characteristic (p = 0.03). Only participation in BreastScreen reading significantly improved all performance metrics: specificity (p = 0.04), sensitivity (p = 0.005), location sensitivity (p < 0.001) and AUC (p < 0.001). Reading > 100 mammograms weekly significantly improved sensitivity (p = 0.03), location sensitivity (p = 0.001), and AUC (p = 0.03).In non-dense breasts, breast fellowship completion significantly improved sensitivity (p = 0.02), location sensitivity (p = 0.04) and AUC (p = 0.002). Participation in BreastScreen reading and reading > 100 mammograms weekly significantly improved only sensitivity (p = 0.002 and p = 0.003, respectively) and location sensitivity (p < 0.001 and p < 0.001, respectively).
CONCLUSION: Participating in screening programs, breast fellowships and reading > 100 mammograms weekly are important indicators of the diagnostic performance of radiologists across dense and non-dense breasts. In dense breasts, optimal performance resulted from participation in a breast screening program.
© The Author(s) 2022.

Entities:  

Keywords:  Mammogram readers; breast cancer; breast density; mammography; observer performance

Year:  2022        PMID: 35111337      PMCID: PMC8801646          DOI: 10.1177/20584601211072279

Source DB:  PubMed          Journal:  Acta Radiol Open


Introduction

Globally, breast cancer is the second most common form of cancer. It is also the second leading cause of cancer-related death among Australian women. Currently, mammography is the only modality considered effective for breast cancer screening as it has been proven to reduce breast cancer mortality, particularly among women aged 50 to 69. Many factors impact the outcome of screening; these include intrinsic limitations of technology, the experience of those interpreting the mammogram, lesion characteristics, and breast density. Researchers often refer to dense breasts as ‘heterogeneously dense’ and ‘extremely dense’ and non-dense breasts as ‘fatty’ and ‘scattered areas of fibroglandular density’. In the United States, it has been estimated that 43% of the screened population has dense breasts. Chinese and Korean women have a higher prevalence of dense breasts than US women, with 49.2% and 54.4%, respectively.[6,7] Data on the Australian population is not available, perhaps due to the lack of breast density notification policy. However, Breast Cancer Network Australia is advocating for breast density policy changes. Increased breast density increases the risk of masking and of cancer. It has also been shown that dense breast composition is directly linked to risks associated with breast cancer, emphasising the need to optimise early detection in women with dense breasts. There is evidence of wide variation in diagnostic efficacy between radiologists or breast image readers. This inter-reader variability requires that intrinsic human factors be considered when designing strategies to improve breast cancer detection. Thus, the impact of radiologists’ characteristics has become a major focus of recent research.[12-14] Several studies have examined the association between reader characteristics such as years of reading mammograms, the number of mammograms read per year, completion of a fellowship in breast imaging and participation in diagnostic workups.[11,13-17] These studies demonstrated wide variation in the relationships between observers’ characteristics and performance in mammography interpretation. However, most published studies assessed the influence of readers’ characteristics with little or no consideration for the impact of breast composition. Thus, the markers of diagnostic efficacy and confidence in dense and non-dense breasts are poorly understood, and further in-depth investigation is needed. Therefore, this study aims to assess the relationship between radiologists’ characteristics and diagnostic performance across dense and non-dense breasts.

Method and materials

Image test sets

Two digital mammography (DM) test sets were developed from a screening population database. Each DM test set contained 60 cases (40 normal and 20 abnormal). The normal cases were confirmed to be normal by at least two radiologists and by a follow-up negative mammogram obtained 2–4 years later. The types and characteristics of the lesions were also established by these radiologists. The abnormal cases contained at least one biopsy-proven cancer lesion. Density classification was determined by a consensus of two consultant breast radiologists with more than 20 years of experience in reading screening mammograms. The cases exhibited a range of breast densities classified as non-dense (≤ 50% glandular tissue) and dense breasts (> 50% glandular tissue). Australia uses the Royal Australian and New Zealand College of Radiologists (RANZCR) synoptic scale, which is similar to the fourth edition BI-RADS Atlas. To make the test set relevant to countries using the RANZCR synoptic scale and BI-RADS fourth edition, breast density of the cases included was classified according to the fourth edition BI-RADS Atlas. The distribution of breast densities was as follows: DM test set 1 included 40% non-dense and 60% dense cases, while DM test set 2 included 45% non-dense and 55% dense cases.

Participants

A total of 128 radiologists specialising in breast imaging were recruited. The mean age of the participants was 53 ± 11.8 years, and their years of reading mammograms ranged from 0.5 to 40 (mean: 13 ± 10.6 years). All participants had completed a training program overseen by the RANZCR. The training program runs for 5 years and includes system-focused rotations in the last 2 years. By the end of their training, radiologists are competent in diagnostic breast imaging, including mammography and ultrasound, and have exposure to breast MRI and to the investigation and staging of metastatic breast cancer. A total of 730 breast cases are studied during the training program (100 diagnostic mammograms, 500 screening mammograms, 100 ultrasounds, 20 MRI and 10 biopsies). The fellowship-trained breast radiologists had an additional 6 months training to specialise in advanced breast imaging and breast procedures. Of the 128 radiologists, 62 (48.4%) read for the national Australian breast screening program (BreastScreen, Australia) at the time of the study, and 42 (32.8%) had completed a 3- to 6-month breast-imaging fellowship. The characteristics of the participants are summarised in Table 1.
Table 1.

Radiologists’ demographic information at the time of completing the DM test set.

Radiologists’ characteristicsDM test set 1DM test set 2Total
Mean age (years)55 (±11.6)48.7 (±11.3)53 (±11.8)
Number of radiologists (M, F)87 (35, 52)41 (19, 22)128 (54, 74)
Breast speciality100%100%100%
Breast screen program readers55.2%34.1%48.4%
Radiologists completed breast fellowship (3 to 6 months)33.3%31.7%32.8%
Mean years reading mammograms14.5 (±10.8)9.6 (±9.3)13 (±10.6)
Number of cases radiologists read weekly
 < 2021 (24%)20 (48.8%)41 (32%)
 20–6017 (19.5%)5 (12.2%)22 (17.2%)
 61–1009 (10.4%)6 (14.6%)15 (11.7%)
 101–15011 (12.7%)1 (2.4%)12 (9.4%)
 151–20016 (18.4%)6 (14.6%)22 (17.2%)
 > 20013 (15%)3 (7.4%)16 (12.5%)
Radiologists’ demographic information at the time of completing the DM test set.

Reading environment

The images were read via the Breast Reader Assessment Strategy (BREAST) platform either at a conference or at different Australian clinical sites using primary displays between 2015 and 2019. Ambient lighting in reading rooms at conferences was set at 15–20 lux to conform with the RANZCR and BreastScreen Australia Accreditation Standards[19,20] as well as with ambient lighting conditions in many clinical settings. Calibrated Barco 5MP medical-grade monochrome liquid crystal display monitors with a resolution of 2049 × 2560 pixels were used.

Study design

Participants completed an electronic survey of their demographic information and work experience - including position, speciality, completion of a fellowship in breast radiology, number of years reading mammography and the number of cases read per week. Subsequently, each reader independently interpreted the images in the test sets and assigned a confidence rating to each decision. If the reader considered the image to be normal, he/she moved to the next case and the case was automatically rated as 1, meaning no cancer was present. If a lesion was detected, the reader marked the lesion’s location and assigned a confidence rating score from 2 to 5, which is compatible with the Tabar/RANZCR classification used in BreastScreen Australia where 2 = benign, 3 = indeterminate/equivocal, 4 = suspicious and 5 = highly suspicious. A rating of 3, 4 or 5 signified malignancy, with higher ratings denoting higher confidence. If a rating of 3 or above was given, the reader was asked to describe the type of breast lesion detected (discrete mass, architectural distortion, spiculated mass, nonspecific density, stellate and calcification) by checking the appropriate box in a pop-up menu. These marks and ratings were then used to assess reader performance.

Statistical analysis

The radiologists’ performances were calculated in terms of specificity, sensitivity, location sensitivity and AUC in dense and non-dense breasts. Location sensitivity was determined by the distance of the mouse click from the breast lesion centre. If distance was not recorded, this indicated that the radiologist marked outside the correct region or did not give any markings. Diagnostic confidence (radiologists’ level of confidence that the detected lesion was malignant) and lesion classification (their ability to correctly classify the lesion into type) in dense and non-dense images were calculated. The diagnostic performance metrics were compared using an independent-samples t-test or a Mann–Whitney U test, depending on the distribution of the data. A chi-squared test (χ2) was conducted to assess the association between radiologists’ characteristics and both diagnostic confidence and lesion classification across dense and non-dense breasts. One-way ANOVA and Kruskal–Wallis tests were applied to compare three independent groups depending on data distribution. p-value ≤ 0.05 was considered statistically significant. These analyses were conducted via Jamovi software (version 1.6.22).

Results

Radiologists’ performances in dense breasts

Table 2 depicts the differences in radiologists’ characteristics and performances with DM for dense breasts. All metrics – including specificity (p = 0.26), sensitivity (p = 0.13), location sensitivity (p = 0.14) and AUC (p = 0.84) – were similar between radiologists who had read mammography for < 10 years, 10–19 years and ≥ 20 years. Specificity (p = 0.004), location sensitivity (p = 0.01) and AUC (p = 0.03), but not sensitivity (p = 0.36), were significantly higher in radiologists who had completed a breast fellowship than in those who had not. Radiologists who work for the BreastScreen Australia program showed significantly higher performance in all metrics – specificity (p = 0.04), sensitivity (p = 0.005), location sensitivity (p < 0.001) and AUC (p < 0.001) – compared to those who do not. Radiologists who read > 100 mammogram cases weekly showed significantly higher sensitivity (p = 0.03), location sensitivity (p = 0.001) and AUC (p = 0.03), but not specificity (p = 0.51).
Table 2.

Comparison of radiologists’ performance characteristics in dense breasts.

Characteristic (readers no.)Specificity (%) p Sensitivity (%) p Location sensitivity (%) p AUC (0–1) p
No. of years reading mammography
 <10 years (63)81.8 ± 13.20.2677.15 ± 19.90.1366.3 ± 20.50.140.794 ± 0.1160.84
 10–19 years (21)75.9 ± 14.583.7 ± 11.371.2 ± 10.40.797 ± 0.079
 10–19 years (21)79.6 ± 14.877.9 ± 15.564.5 ± 17.20.784 ± 0.107
Breast fellowship (3–6 months)
 Yes (42)87 (78.91)* 0.004 80.5 ± 16.80.3672.4 ± 16.5 0.01 0.820 ± 0.103 0.03
 No (86)81 (66.90)*77.5 ± 17.663.7 ± 18.20.777 ± 0.107
BreastScreen Australia program reader
 Yes (62)82.7 ± 13.1 0.04 84.6 (77.92)* 0.005 77 (69.83)*< 0.0010.828 ± 0.108< 0.001
 No (66)77.6 ± 14.683.3 (62.85)*67 (54.75)*0.757 ± 0.093
No. of cases read per week
 ≤ 100 cases (50)81 ± 15.20.5185 (76.92)* 0.03 76 (67.83)* 0.001 0.817 ± 0.091 0.03
 ≤ 100 cases (78)79.4 ± 13.383 (67.92)*68 (54.68)*0.775 ± 0.113

(*) signifies median values, including 1st and 3rd quartiles, where significant values resulted from the Mann–Whitney U test. Bold values indicate statistical significance at the p-value ≤ 0.05 level.

Comparison of radiologists’ performance characteristics in dense breasts. (*) signifies median values, including 1st and 3rd quartiles, where significant values resulted from the Mann–Whitney U test. Bold values indicate statistical significance at the p-value ≤ 0.05 level.

Radiologists’ performances in non-dense breasts

Table 3 shows the differences in radiologists’ characteristics and performances with DM for non-dense breasts. All metrics, including specificity (p = 0.08), sensitivity (p = 0.44), location sensitivity (p = 0.09) and AUC (p = 0.25), were similar between radiologists who had read mammography for < 10 years, 10–19 years and ≥ 20 years. Specificity (p = 0.13) was similar between radiologists who had completed the breast fellowship and those who had not; however, sensitivity (p = 0.02), location sensitivity (p = 0.4) and AUC (p = 0.002) were significantly higher among those who had completed the breast fellowship. Radiologists who work for the BreastScreen Australia program showed similar specificity (p = 0.60) and AUC (p = 0.09), but significantly higher sensitivity (p = 0.002) and location sensitivity (p < 0.001), compared to those who do not. Radiologists who read > 100 mammogram cases weekly showed significantly higher sensitivity (p = 0.003) and location sensitivity (p < 0.001), but not specificity (p = 0.18) or AUC (p = 0.35).
Table 3.

Comparison of radiologists’ performance characteristics for non-dense breasts.

Characteristics (readers no.)Specificity (%) p Sensitivity (%) p Location sensitivity (%) p AUC (0–1) p
No. of years reading mammography
 < 10 years (63)80 ± 15.430.0878.7 ± 17.50.4465.2 ± 220.090.797 ± 0.1040.25
 10–19 years (21)78.44 ± 15.6383.6 ± 15.275.25 ± 16.70.809 ± 0.105
 ≥ 20 years (44)73 ± 15.981.5 ± 16.670.9 ± 16.80.769 ± 0.102
Breast fellowship (3–6 months)
 Yes (42)80.3 ± 15.60.1385 ± 16.3 0.02 74 ± 18.3 0.04 0.829 ± 0.104 0.002
 No (86)75.8 ± 15.877.9 ± 16.566.5 ± 20.20.769 ± 0.098
BreastScreen Australia program reader
 Yes (62)76.6 ± 17.40.6085 ± 14.3 0.002 71.4 (71.86)*< 0.0010.805 ± 0.1050.09
 No (66)78 ± 14.175.8 ± 17.762.5 (43.83)*0.774 ± 0.101
No. of cases read per week
 >100 cases (50)75 ± 180.1885.7 ± 14.5 0.003 73 (71.86)*< 0.0010.800 ± 0.1070.35
 ≤100 cases (78)88 ± 14.176.8 ± 17.271 (43.83)*0.782 ± 0.102

(*) signifies median values, including 1st and 3rd quartiles, where significant values resulted from the Mann–Whitney U test. Bold values indicate statistical significance at the p-value ≤ 0.05 level.

Comparison of radiologists’ performance characteristics for non-dense breasts. (*) signifies median values, including 1st and 3rd quartiles, where significant values resulted from the Mann–Whitney U test. Bold values indicate statistical significance at the p-value ≤ 0.05 level.

Association between radiologists’ characteristics and diagnostic confidence when reporting breast cancer across breast densities

Diagnostic confidence among radiologists who had read mammography for < 10 years, 10–19 years and ≥ 20 years did not differ significantly in dense (p = 0.42) or non-dense (p = 0.99) breasts. Radiologists who had completed a breast fellowship showed no significant increase in diagnostic confidence compared to those who had not in either dense (p = 0.06) or non-dense (p = 0.08) breasts. BreastScreen program readers showed significantly lower confidence levels (i.e. used more scores of 3 (indeterminate/equivocal)) than radiologists who do not read for the program (dense: p = 0.008; non-dense p < 0.001). Radiologists who read a higher volume (> 100) of mammogram cases weekly also had significantly lower confidence levels than those who read ≤ 100 mammogram cases per week when reporting cancer in dense (p = 0.004) and non-dense (p < 0.001) breasts, as shown in Table 4.
Table 4.

Association between radiologists’ characteristics and diagnostic confidence when reporting breast cancer.

Reader characteristics (readers no.)Confidence score (dense cases)Confidence score (non-dense cases)
4 or 53 p 4 or 53 p
No. of years reading mammography
 < 10 years (63)36.4% (289/794)63.6% (505/794)0.4234.12% (159/466)65.88% (307/466)0.99
 10–19 years (21)35% (92/266)65% (174/266)35.7% (55/154)64.3% (99/154)
 ≥ 20 years (44)40% (225/563)60% (338/563)34.9% (111/318)65.1% (207/318)
Breast fellowship (3–6 months)
 Yes (42)44% (187/425)56% (241/428)0.0638.5% (101/262)61.5% (161/262)0.08
 No (86)49.5% (418/845)50.5% (427/845)45.4% (224/493)54.6% (269/493)
BreastScreen Australia program reader
 Yes (62)44% (292/664)56% (372/664) 0.008 36% (137/383)64.2% (246/383)< 0.001
 No (66)51.6% (314/609)48.4% (295/609)50.5% (188/372)49.5% (184/372)
No. of cases read per week
 > 100 cases (50)42.8% (229/535)57.2% (306/535) 0.004 34.4% (107/311)65.6% (204/311)< 0.001
 ≤ 100 cases (78)51.1% (377/738)48.9% (361/738)49.3% (219/444)50.7% (225/444)

Score 3 (indeterminate/equivocal); score 4 (suspicious); score 5 (highly suspicious). Bold values indicate statistical significance at the p-value ≤ 0.05 level.

Association between radiologists’ characteristics and diagnostic confidence when reporting breast cancer. Score 3 (indeterminate/equivocal); score 4 (suspicious); score 5 (highly suspicious). Bold values indicate statistical significance at the p-value ≤ 0.05 level.

Association between radiologists’ characteristics and their ability to classify lesions into types across breast densities

When classifying breast lesions, the performance of radiologists who had read mammography for < 10 years, 10–19 years and ≥ 20 years did not differ significantly in dense (p = 0.56) or non-dense (p = 0.96) breasts. Completion of the breast fellowship did not impact radiologists’ lesion classification performances in dense (p = 0.16) or non-dense (p = 0.44) cases. However, in dense cases (p = 0.03), BreastScreen program readers performed significantly better than those who did not read for the program; this was not true in non-dense cases (p = 0.12). Also, radiologists who interpreted > 100 mammogram cases weekly more accurately classified breast lesions in dense breasts (p = 0.03) but not in non-dense breasts (p = 0.17), as shown in Table 5.
Table 5.

Association between radiologists’ characteristics and lesion classification.

Reader characteristics (readers no.)Lesion classification into types (dense cases)Lesion classification into types (non-dense cases)
TrueFalse p TrueFalse p
No. of years reading mammography
 < 10 years (63)30% (238/794)70% (556/794)0.5641.2% (192/466)58.8% (274/466)0.96
 10–19 years (21)33% (88/266)67% (178/266)43.5% (67/154)56.5% (87/154)
 ≥ 20 years (44)33% (185/563)67% (378/563)35% (111/318)65% (207/318)
Breast fellowship (3–6 months)
 Yes (42)43% (184/428)57% (244/428)0.1654.6% (143/262)45.4% (119/262)0.44
 No (86)38.7% (327/845)61.3% (518/845)51.3% (253/493)48.7% (240/493)
BreastScreen Australia program reader
 Yes (62)43% (286/664)57% (378/664) 0.03 55.4% (212/383)44.6% (171/383)0.12
 No (66)37% (225/609)63% (384/609)49.5% (184/372)50.5% (188/372)
No. of cases read per week
 > 100 cases (50)44% (234/535)56% (301/535) 0.03 55.6% (173/311)44.4% (138/311)0.17
 ≤ 100 cases (78)37.5% (277/738)62.5% (461/738)50.2% (223/444)49.8% (221/444)

Lesion types: Stellate, discrete mass, spiculated mass, nonspecific density, calcification and architectural distortion. Bold values indicate statistical significance at the p-value ≤ 0.05 level.

Association between radiologists’ characteristics and lesion classification. Lesion types: Stellate, discrete mass, spiculated mass, nonspecific density, calcification and architectural distortion. Bold values indicate statistical significance at the p-value ≤ 0.05 level.

Discussion

This observational study was conducted to assess radiologists’ performances and markers of good performance across different breast compositions. The evidence suggests a difference in radiologists’ performance characteristics across dense and non-dense breasts, indicating that participation in screening programmes, completing breast fellowship training and reading >100 mammograms weekly are important diagnostic performance indicators, but not the number of years reading mammography. The literature shows wide variation in the search, perception and decision-making abilities of radiologists that are concomitant with differences in performance in the interpretation of mammographic images, suggesting that human limitations significantly impact the efficacy of screening mammography. Differences in reader ability and interaction with radiological images cannot be completely mitigated, but should be exploited to improve diagnostic efficacy. We thought that in a simulated clinical environment, readers may exhibit a trade-off between sensitivity and specificity in dealing with suspicious cases. However, BreastScreen readers and those with breast fellowship training were not influenced by this trade-off, as demonstrated by the significantly higher specificity compared to other characteristics, particularly in dense breasts. Whilst the impact of fellowship training and volume read on performance is consistent with most published literature,[11,17,22,23] it is unclear whether the significantly higher performance observed in BreastScreen readers is due to feedback from the service. It is possible that these three factors are interconnected and work together to improve performance. The lack of association between years of reading mammograms and performance is reasonable and consistent with data from both data linkage and observer performance studies.[12,13,17,24] Our findings suggest that years reading mammograms may not necessarily capture experience because factors such as mentorship, participation in diagnostic workup, feedback and interactions with images of different disease presentations may influence how radiologists build image interpretation skills. Diagnostic confidence is also an important factor of radiologists’ performance, as it is associated with greater accuracy in detecting breast cancer and influences decision-making regarding recall for further assessment or biopsy.[22,25] However, diagnostic confidence is complex, involving visual perceptions and clinical judgements that depend on other factors, such as image quality and the interpreters’ capabilities. Interestingly, BreastScreen Australia and high-volume readers demonstrated significantly lower diagnostic confidence, using a score of 3 when reporting breast cancer across dense and non-dense breasts. This is consistent with the results of a previous observational study that did not consider mammographic density. This could be due to a cohort of more risk-averse readers who do not want to overemphasise the significance of the lesion or who assume that mammograms alone are unreliable and require additional information from supplemental imaging and biopsy for confirmation. In the Tabar/RANZCR scoring scheme, a score of 3 indicates that the lesion requires further investigation, usually through percutaneous needle biopsy. Diagnostic confidence based on the assessment categories is very subjective; the decision to recall or not is the key parameter, as any score of 3 or above will result in a recall. These findings have a few implications for policy and practice. First, radiologists’ characteristics associated with performance in dense and non-dense breasts can be used to optimise pairing strategies in countries such as Australia, where independent double reading of mammograms is practiced. This strategy may increase the chance that if a lesion is missed by one radiologist due to breast composition, it will be detected by another. Second, these findings can be used to identify radiologists who may benefit from tailored training in identifying breast cancer in different density breasts and inform educational interventions to improve their performance. Third, radiological lesion classification into types has been associated with specific histological findings, and variation in the classification of lesions could affect further assessment and patient management. Therefore, familiarity with lesion features may eliminate some intrinsic errors associated with radiologists’ diagnostic performance and decisions, and test-set data may be useful for training to improve radiologists’ ability to detect and classify malignant features on mammograms. These findings suggest the need for observational studies exploring the impact of knowledge of mammographic lesion features on breast cancer detection and confidence levels across dense and non-dense breasts. This study is not without limitations. First, the number of dense cases was comparable to that of non-dense cases, and such weighting may not reflect real screening populations. However, the number of cases needed to be similarly distributed to avoid selection bias. This is supported by a previous study that found readers show significantly higher sensitivity rates in dense breasts when fewer dense breast cases are included in the dataset. Second, these findings may not completely reflect the performance of this cohort of radiologists in actual screening practice because interval cancers were not considered. However, a previous study comparing the performance of the same cohort of Australian radiologists in both clinical and test settings showed no difference in performance, suggesting that test set data can reasonably predict performance in a clinical setting. Clinical audits to assess radiologists’ screening performance require several years of follow-up to establish true interval cancers and negative mammograms, and the results of these audits are provided to the clinical practice rather than the individual radiologist. Therefore, test set data may provide opportunities to establish reader characteristics associated with performance across breasts of different compositions and feedback for individual radiologists. Third, it is possible that not all testing sites conformed to the RANZCR and BreastScreen Australia Accreditation Standards ambient lighting standards, although such differences should have a negligible impact on the findings. To our knowledge, no study has closely examined the characteristics of BreastScreen Australia readers associated with improved performance in different breast compositions. Therefore, our study provides baseline data to optimise diagnostic efficacy and confidence in dense breasts. In conclusion, participating in a screening program reading, fellowship in breast imaging and weekly volumes read of greater than 100 mammogram cases are the most important indicators of diagnostic performance across dense and non-dense breasts. In dense breasts, optimal performance was demonstrated by screening program readers. These findings have practical implications for helping breast screening programs achieve better outcomes in dense and non-dense breasts.
  23 in total

1.  Malignancy detection in digital mammograms: important reader characteristics and required case numbers.

Authors:  Warren M Reed; Warwick B Lee; Jennifer N Cawson; Patrick C Brennan
Journal:  Acad Radiol       Date:  2010-08-17       Impact factor: 3.173

2.  Markers of good performance in mammography depend on number of annual readings.

Authors:  Mohammad A Rawashdeh; Warwick B Lee; Roger M Bourne; Elaine A Ryan; Mariusz W Pietrzyk; Warren M Reed; Robert C Heard; Deborah A Black; Patrick C Brennan
Journal:  Radiology       Date:  2013-06-04       Impact factor: 11.105

3.  Number of mammography cases read per year is a strong predictor of sensitivity.

Authors:  Wasfi I Suleiman; Sarah J Lewis; Dianne Georgian-Smith; Michael G Evanoff; Mark F McEntee
Journal:  J Med Imaging (Bellingham)       Date:  2014-05-07

4.  The "laboratory" effect: comparing radiologists' performance and variability during prospective clinical and laboratory mammography interpretations.

Authors:  David Gur; Andriy I Bandos; Cathy S Cohen; Christiane M Hakim; Lara A Hardesty; Marie A Ganott; Ronald L Perrin; William R Poller; Ratan Shah; Jules H Sumkin; Luisa P Wallace; Howard E Rockette
Journal:  Radiology       Date:  2008-08-05       Impact factor: 11.105

5.  The influence of increased ambient lighting on mass detection in mammograms.

Authors:  Benjamin J Pollard; Ehsan Samei; Amarpreet S Chawla; Jay Baker; Sujata Ghate; Connie Kim; Mary S Soo; Noriyuki Hashimoto
Journal:  Acad Radiol       Date:  2009-03       Impact factor: 3.173

6.  Prevalence of mammographically dense breasts in the United States.

Authors:  Brian L Sprague; Ronald E Gangnon; Veronica Burt; Amy Trentham-Dietz; John M Hampton; Robert D Wellman; Karla Kerlikowske; Diana L Miglioretti
Journal:  J Natl Cancer Inst       Date:  2014-09-12       Impact factor: 13.506

7.  Cancer incidence and mortality worldwide: sources, methods and major patterns in GLOBOCAN 2012.

Authors:  Jacques Ferlay; Isabelle Soerjomataram; Rajesh Dikshit; Sultan Eser; Colin Mathers; Marise Rebelo; Donald Maxwell Parkin; David Forman; Freddie Bray
Journal:  Int J Cancer       Date:  2014-10-09       Impact factor: 7.396

8.  Prevalence of Women with Dense Breasts in Korea: Results from a Nationwide Cross-sectional Study.

Authors:  Hye-Mi Jo; Eun Hye Lee; Kyungran Ko; Bong Joo Kang; Joo Hee Cha; Ann Yi; Hae Kyoung Jung; Jae Kwan Jun
Journal:  Cancer Res Treat       Date:  2019-01-29       Impact factor: 4.679

9.  Association between radiologists' experience and accuracy in interpreting screening mammograms.

Authors:  Eduard Molins; Francesc Macià; Francesc Ferrer; Maria-Teresa Maristany; Xavier Castells
Journal:  BMC Health Serv Res       Date:  2008-04-25       Impact factor: 2.655

Review 10.  Errors in Mammography Cannot be Solved Through Technology Alone

Authors:  Ernest Usang Ekpo; Maram Alakhras; Patrick Brennan
Journal:  Asian Pac J Cancer Prev       Date:  2018-02-26
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.