| Literature DB >> 35054348 |
Anna Landsmann1, Jann Wieler1, Patryk Hejduk1, Alexander Ciritsis1, Karol Borkowski1, Cristina Rossi1, Andreas Boss1.
Abstract
The aim of this study was to investigate the potential of a machine learning algorithm to accurately classify parenchymal density in spiral breast-CT (BCT), using a deep convolutional neural network (dCNN). In this retrospectively designed study, 634 examinations of 317 patients were included. After image selection and preparation, 5589 images from 634 different BCT examinations were sorted by a four-level density scale, ranging from A to D, using ACR BI-RADS-like criteria. Subsequently four different dCNN models (differences in optimizer and spatial resolution) were trained (70% of data), validated (20%) and tested on a "real-world" dataset (10%). Moreover, dCNN accuracy was compared to a human readout. The overall performance of the model with lowest resolution of input data was highest, reaching an accuracy on the "real-world" dataset of 85.8%. The intra-class correlation of the dCNN and the two readers was almost perfect (0.92) and kappa values between both readers and the dCNN were substantial (0.71-0.76). Moreover, the diagnostic performance between the readers and the dCNN showed very good correspondence with an AUC of 0.89. Artificial Intelligence in the form of a dCNN can be used for standardized, observer-independent and reliable classification of parenchymal density in a BCT examination.Entities:
Keywords: artificial intelligence; machine learning; photon counting breast-CT; spiral breast-CT
Year: 2022 PMID: 35054348 PMCID: PMC8775263 DOI: 10.3390/diagnostics12010181
Source DB: PubMed Journal: Diagnostics (Basel) ISSN: 2075-4418
Figure 1Spiral breast-CT density classification with one example for each category. (A). Partial or complete involution with every lesion visible; (B). Scattered glandular tissue with lesions larger than 10 mm conclusively visible; (C). Heterogenous dense glandular tissue with lesions of 10 mm potentially not visible; (D). Very dense tissue with restricted visibility of lesions. Raw data images with 0.3 mm slice thickness, coronal plane.
Characteristics for each Model of the Deep Convolutional Neural Network.
| Model 1 | Model 2 | Model 3 | Model 4 | |
|---|---|---|---|---|
|
| brayZNet | brayZNet | brayZNet | brayZNet |
|
| 1 × 10−5 | 1 × 10−5 | 1 × 10−5 | 1 × 10−5 |
|
| Cross entropy | Cross entropy | Cross entropy | Cross entropy |
|
| Adam | Adam | SGD | Adam |
|
| {‘zooming’: 0.1, ‘rotation’: 45.0, ‘horizontal_shift’: 0.1, ‘vertical_shift’: 0.1, ‘brightness’: 0.0} | {‘zooming’: 0.1, ‘rotation’: 45.0, ‘horizontal_shift’: 0.1, ‘vertical_shift’: 0.1, ‘brightness’: 0.0} | {‘zooming’: 0.1, ‘rotation’: 45.0, ‘horizontal_shift’: 0.1, ‘vertical_shift’: 0.1, ‘brightness’: 0.0} | {‘zooming’: 0.1, ‘rotation’: 45.0, ‘horizontal_shift’: 0.1, ‘vertical_shift’: 0.1, ‘brightness’: 0.0} |
|
| 160 | 160 | 160 | 160 |
|
| 8 | 8 | 8 | 8 |
|
| 0.5 | 0.5 | 0.5 | 0.5 |
|
| [512, 512, 1] | [256, 256, 1] | [512, 512, 1] | [512, 512, 1] |
|
| None | None | None | [0.12826739057573872, 0.8474049572056288, 0.200998651126856, 0.8363919170216573] |
|
| 2 | 2 | 2 | 2 |
|
| 128 | 128 | 128 | 128 |
|
| l1 = 1 × 10−6, l2 = 1 × 10−6 | l1 = 1 × 10−6, l2 = 1 × 10−6 | l1 = 1 × 10−6, l2 = 1 × 10−6 | l1 = 1 × 10−6, l2 = 1 × 10−6 |
|
| 0.8041666746139526 | 0.8583333492279053 | 0.7354166507720947 | 0.8020833134651184 |
Patient overview regarding breast density distribution and ultrasound examinations, with n describing the total number of patients.
| Density Level | Ultrasound (US) | Reason for US Examination | |||
|---|---|---|---|---|---|
| Yes | No | Density | Other | ||
|
| 58 | 14 | 44 | 0 | 14 |
|
| 118 | 64 | 54 | 28 | 36 |
|
| 83 | 80 | 3 | 57 | 23 |
|
| 58 | 56 | 2 | 43 | 13 |
|
| 317 | 214 | 103 | 128 | 86 |
Figure 2Training and validation accuracy and loss curves for the 4 different deep learning configurations. Model 1: Adam optimizer, Model 2: low resolution, Model 3: SGD optimizer and Model 4: cropped image. Highest accuracy was reached by Model 2.
Confusion matrices of the “real-world” dataset compared to the assessment of the radiologist as the ground-truth. Numbers in bold highlight the correctly assessed images.
| Predicted Density Level | |||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Model 1 | Model 2 | Model 3 | Model 4 | ||||||||||||||
| A | B | C | D | A | B | C | D | A | B | C | D | A | B | C | D | ||
|
|
|
| 15 | 0 | 0 |
| 15 | 1 | 0 |
| 13 | 0 | 0 |
| 43 | 0 | 0 |
|
| 19 |
| 23 | 0 | 12 |
| 19 | 0 | 31 |
| 10 | 3 | 4 |
| 15 | 0 | |
|
| 0 | 25 |
| 2 | 0 | 11 |
| 5 | 1 | 38 |
| 1 | 0 | 20 |
| 2 | |
|
| 0 | 0 | 10 |
| 0 | 1 | 4 |
| 0 | 1 | 29 |
| 0 | 0 | 11 |
| |
Assessment of breast-CT density of two readers and dCNN Model 2 in 60 images.
| dCNN | Reader 1 | Reader 2 | ||
|---|---|---|---|---|
|
| 14 | 15 | 16 |
|
|
| 16 | 18 | 19 | |
|
| 15 | 17 | 10 |
|
|
| 15 | 10 | 15 |
Figure 3Example dCNN evaluations for breast-CT images showing different density levels, one example for each category.
Kappa values between each pair of both readers, “ground-truth” and dCNN Model 2 using a four-level density scale. According to Landis and Koch: 0.6 to 0.8 substantial and >0.8 almost perfect agreement.
| Ground-Truth | dCNN | Reader 1 | Reader 2 | |
|---|---|---|---|---|
|
| 0.84 | 0.87 | 0.82 | |
|
| 0.71 | 0.73 | ||
|
| 0.73 | |||
|
|
Figure 4Receiver Operating Characteristics (ROC) Analysis for the Human Readout compared to the dCNN (Model 2) using a binary classification system: 0 = lower density (level A and B), 1 = high density (level C and D). Diagnostic Accuracy is expressed as the area under the curve (AUC).