| Literature DB >> 35366123 |
Lorenz A Kapsner1,2, Sabine Ohlmeyer3, Lukas Folle4, Frederik B Laun3, Armin M Nagel3, Andrzej Liebert3, Hannes Schreiter3, Matthias W Beckmann5, Michael Uder3, Evelyn Wenkel3, Sebastian Bickelhaupt3,6.
Abstract
OBJECTIVES: To automatically detect MRI artifacts on dynamic contrast-enhanced (DCE) maximum intensity projections (MIPs) of the breast using deep learning.Entities:
Keywords: Artifacts; Breast; Contrast agent; Magnetic resonance imaging; Neural network
Mesh:
Substances:
Year: 2022 PMID: 35366123 PMCID: PMC9381479 DOI: 10.1007/s00330-022-08626-5
Source DB: PubMed Journal: Eur Radiol ISSN: 0938-7994 Impact factor: 7.034
Demographic data, sample characteristics, and target class distribution across the training dataset and test dataset. IQR interquartile range
| Variable | Overall sample | Training dataset | Test dataset |
|---|---|---|---|
| 1794 | 1378 | 416 | |
| Age | |||
| Median age (IQR) (years) | 49 (16) | 49 (16) | 50 (18) |
| Median age (IQR) at first acquisition (years) | 50 (17) | 50 (16.75) | 51 (18) |
| 2265 | 1827 | 438 | |
| One examination | 1461 | 1067 | 394 |
| Two examinations | 225 | 203 | 22 |
| Three examinations | 80 | 80 | |
| Four examinations | 27 | 27 | |
| Six examinations | 1 | 1 | |
| 4530 | 3654 | 876 | |
| Left breast | 2265 | 1827 | 438 |
| Right breast | 2265 | 1827 | 438 |
| 2332 (51%) | 1951 (53%) | 381 (43%) | |
| Left breast | 1147 (51%) | 959 (52%) | 188 (43%) |
| Right breast | 1185 (52%) | 992 (54%) | 193 (44%) |
Ensemble classifier performance on the holdout test dataset. The table shows the performance of the ensemble classifiers for ResNet and DenseNet on the holdout test dataset. AUROC area under the receiver operating characteristic curve, AUPRC area under the precision-recall curve, PPV positive predictive value, NPV negative predictive value
| Variable | DenseNet | ResNet |
|---|---|---|
| Accuracy | 0.858 | 0.848 |
| AUROC | 0.940 | 0.923 |
| AUPRC | 0.928 | 0.907 |
| Sensitivity | 0.900 | 0.840 |
| Specificity | 0.826 | 0.855 |
| PPV | 0.800 | 0.816 |
| NPV | 0.915 | 0.874 |
Cross-validation results. The table shows the performance measures of the 5 cross-validation models for ResNet and DenseNet on their test datasets. CV cross-validation, Mean (unweighted) average over the 5 CV folds, SD (unweighted) standard deviation over the 5 CV folds, AUROC area under the receiver operating characteristic curve, AUPRC area under the precision-recall curve, PPV positive predictive value, NPV negative predictive value, sec seconds, Time per epoch average time per epoch observed until convergence
| Model | Variable | CV fold 1 | CV fold 2 | CV fold 3 | CV fold 4 | CV fold 5 | Mean (SD) |
|---|---|---|---|---|---|---|---|
| DenseNet | Best epoch | 177 | 146 | 101 | 190 | 181 | 159.0 (± 36.407) |
| Time per epoch (sec) | 17.740 | 17.738 | 17.765 | 17.732 | 17.711 | 17.737 (± 0.019) | |
| Accuracy | 0.819 | 0.844 | 0.798 | 0.803 | 0.792 | 0.811 (± 0.021) | |
| AUROC | 0.906 | 0.906 | 0.886 | 0.901 | 0.881 | 0.896 (± 0.012) | |
| AUPRC | 0.925 | 0.921 | 0.894 | 0.923 | 0.894 | 0.911 (± 0.016) | |
| Sensitivity | 0.797 | 0.838 | 0.831 | 0.803 | 0.800 | 0.814 (± 0.019) | |
| Specificity | 0.845 | 0.850 | 0.760 | 0.803 | 0.782 | 0.808 (± 0.039) | |
| PPV | 0.854 | 0.865 | 0.798 | 0.824 | 0.808 | 0.830 (± 0.029) | |
| NPV | 0.785 | 0.822 | 0.797 | 0.780 | 0.773 | 0.791 (± 0.019) | |
| ResNet | Best epoch | 67 | 82 | 64 | 118 | 113 | 88.800 (± 25.371) |
| Time per epoch (sec) | 8.162 | 8.152 | 8.195 | 8.125 | 8.129 | 8.153 (± 0.028) | |
| Accuracy | 0.807 | 0.813 | 0.795 | 0.796 | 0.789 | 0.800 (± 0.010) | |
| AUROC | 0.884 | 0.891 | 0.867 | 0.884 | 0.871 | 0.879 (± 0.010) | |
| AUPRC | 0.905 | 0.913 | 0.889 | 0.913 | 0.887 | 0.901 (± 0.013) | |
| Sensitivity | 0.779 | 0.818 | 0.805 | 0.780 | 0.805 | 0.797 (± 0.017) | |
| Specificity | 0.839 | 0.806 | 0.783 | 0.815 | 0.771 | 0.803 (± 0.027) | |
| PPV | 0.847 | 0.829 | 0.809 | 0.829 | 0.801 | 0.823 (± 0.018) | |
| NPV | 0.769 | 0.795 | 0.778 | 0.763 | 0.775 | 0.776 (± 0.012) |
Comparison of the DenseNet and the ResNet network architectures. The table shows the p values computed with the Wilcoxon rank sum test to compare the performance results of the two utilized network architectures during cross-validation. AUROC area under the receiver operating characteristic curve, AUPRC area under the precision-recall curve, PPV positive predictive value, NPV negative predictive value, sec seconds, Time per epoch average time per epoch observed until convergence, * p value < 0.05, ** p value < 0.01
| Variable | |
|---|---|
| Best epoch | 0.032* |
| Time per epoch (sec) | 0.008** |
| Accuracy | 0.421 |
| AUROC | 0.093 |
| AUPRC | 0.207 |
| Sensitivity | 0.530 |
| Specificity | 1.000 |
| PPV | 1.000 |
| NPV | 0.151 |
Fig. 1ResNet plots. The figure shows the receiver operating characteristic (ROC) curve (left column) and the precision-recall curve (mid column) and the loss curves (right column) for the ResNet architecture. Row 1: ROC and PR curve averaged over 5 cross-validation folds. Row 2: ROC and PR curve for the ensemble’s prediction on the independent holdout test dataset. The training loss (dark blue) and the validation loss (yellow) curves are averaged over 5 CV folds
Fig. 2DenseNet plots. The figure shows the receiver operating characteristic (ROC) curve (left column) and the precision-recall curve (mid column) and the loss curves (right column) for the DenseNet architecture. Row 1: ROC and PR curve averaged over 5 cross-validation folds. Row 2: ROC and PR curve for the ensemble’s prediction on the independent holdout test dataset. The training loss (dark blue) and the validation loss (yellow) curves are averaged over 5 CV folds
Fig. 3Class activation maps (examples). The figure shows the Grad-CAM++ visualizations for one each of a true positive (TP), true negative (TN), false positive (FP), and false negative (FN) predicted images from the holdout test dataset for the respective predicted class. The heatmaps depict with the color gradient from blue to red the relevance of each pixel for the inference of the respective class.
Fig. 4BI-RADS 5 and BI-RADS 6 lesions in clinical cases (examples). Each tile of the figure presents the left or right breast of one clinical case with a diagnosed BI-RADS 5 or BI-RADS 6 lesion. Row 1 (a–c) shows images without the presence of artifacts (a: BI-RADS 6; b: BI-RADS 6; c: BI-RADS 5). Row 2 (d–f) shows images that contain artifacts with no or moderate influence on the diagnostic assessment (d: BI-RADS 5; d: BI-RADS 5; f: BI-RADS 6). Row 3 (g–i) shows images with artifacts potentially impeding the diagnostic evaluation (g: BI-RADS 6; h: BI-RADS 6; i: BI-RADS 5)