| Literature DB >> 32934610 |
Anke Meyer-Bäse1, Lia Morra2, Uwe Meyer-Bäse3, Katja Pinker4,5.
Abstract
Recent advances in artificial intelligence (AI) and deep learning (DL) have impacted many scientific fields including biomedical maging. Magnetic resonance imaging (MRI) is a well-established method in breast imaging with several indications including screening, staging, and therapy monitoring. The rapid development and subsequent implementation of AI into clinical breast MRI has the potential to affect clinical decision-making, guide treatment selection, and improve patient outcomes. The goal of this review is to provide a comprehensive picture of the current status and future perspectives of AI in breast MRI. We will review DL applications and compare them to standard data-driven techniques. We will emphasize the important aspect of developing quantitative imaging biomarkers for precision medicine and the potential of breast MRI and DL in this context. Finally, we will discuss future challenges of DL applications for breast MRI and an AI-augmented clinical decision strategy.Entities:
Year: 2020 PMID: 32934610 PMCID: PMC7474774 DOI: 10.1155/2020/6805710
Source DB: PubMed Journal: Contrast Media Mol Imaging ISSN: 1555-4309 Impact factor: 3.161
Brief overview of common data-driven techniques used in breast MRI.
| Technique | Advantages | Disadvantages | References |
|---|---|---|---|
|
| |||
| Ensemble of decision trees | Decision using branches | Prone to overfitting | [ |
| [ | |||
| Random forest | High performance | Prone to overfitting | [ |
| [ | |||
| Support vector machines | Transforms nonlinear classification problem into linear one | Difficult computation in high-dimensional data space | [ |
| [ | |||
| [ | |||
| Neural networks | Weights need to be adapted for training | No strategy to determine network structure | [ |
| [ | |||
| [ | |||
| Deep learning | State-of-the-art in image-derived features | Computationally intensive | [ |
| [ | |||
| [ | |||
|
| |||
|
| |||
| Clustering ( | Brief training duration | Number of clusters must be known in advance | [ |
| Topological data analysis | Interpretable data mapping | Divided clusters due to mapping | [ |
Figure 1Differences between conventional and deep learning in breast MRI for the lesion discrimination task. The upper part of the image represents the traditional radiomic-based processing. Features such as texture, shape, and histogram are fused to describe the tumor. These engineered features are defined based on expert knowledge. They are extracted from an accurate segmentation which may be performed automatically or, more often, in a semiautomatic fashion by an expert radiologist. The lower part shows the DL-based processing. Several deeper layer features from low level (edges) to high level (objects) are automatically learned by the network. This approach does not require an explicit segmentation step and can be directly applied to the raw images, trained only from lesion-level class labels.
Figure 2Flowchart shows selection of studies for inclusion in the narrative review (a); selected studies are further characterized according to the main focus (b).
Segmentation applications in breast MRI.
| DL technique | Evaluation results | Used dataset | Reference |
|---|---|---|---|
| 2D U-net applied slice-by-slice | Dice = 95.90 ± 0.74 | 42 patients DCE-MRI | [ |
| 3TP U-net | Dice = 61 ± 11.84 | 35 DCE-MRI 4D data | [ |
| GOCS-DLP shape prior based on semantic segmentation based on DL | Dice = 77 ± 13 | 117 patients DCE-MRI, T2- and T1-weighted images | [ |
| 2D U-net applied slice-by-slice | Dice = 97 | 50 DCE-MR images | [ |
| Hierarchical multistage U-net with dice loss | Dice = 72 ± 24 | Training set: 224 DCE-MRI cases; test set: 48 DCE-MRI cases | [ |
| Comparison of 2D U-net and 2D SegNet models with transfer learning from DCE-MRI to DWI | Dice = 72 ± 16 | Training: 39 DCE-MR cases and 15 DWI-MR cases; testing: 10 representative DWI-MR slices | [ |
| 2D U-net applied slice-by-slice to multiplanar sections followed by voxel-level fusion | Dice = 96 ± 0.3 | Training: 42 + 88 T1-weighted MRI series (10-fold cross-validation) | [ |
The most common performance measures are the Dice coefficient and the by-voxel accuracy (ACC), sensitivity (Sn), and specificity (Sp). All performance values reported are percentages.
Figure 3Deep learning network with U-net architecture. Reprinted with permission from [34].
Figure 4Two different approaches for applying U-net to breast and fibroglandular tissue (FGT) segmentation. The upper figure shows 2C U-nets, where two consecutive U-nets are used. The figure below illustrates the other approach, a single U-net with 3-class outputs. Pnb, Pbreast, Pfat, and PFGT denote the probability values of voxels to belong to nonbreast, breast, fat, and FGT, respectively. Reprinted with permission from [34].
Detection of breast lesions in breast MRI using DL.
| DL technique | Evaluation results | Dataset | References |
|---|---|---|---|
| Model agnostic saliency | TPR = 80 | 117 subjects | [ |
| U-net | Acc = 94.2 | 67 MR images T1W, T2W, DWI, and DCE-MRI | [ |
| Patch-based analysis with ResNet50 backbone | AUC = 0.817 | 335 MR images of 17 different histological subtypes | [ |
| Deep Q-network | Sn = 80 | 117 DCE-MR and T1-weighted images | [ |
| Unsupervised saliency analysis and CNN | Acc = 86 ± 2 | 193 DCE-MR images | [ |
| Two-level U-net and dual-stream CNN | CPM = 64.29 | Training: 201 DCE-MR images | [ |
Selected studies reporting classification of breast lesions in breast MRI using DL.
| DL technique | Evaluation results | Dataset | Reference |
|---|---|---|---|
| 3D CNN from scratch | AUC = 0.739 (2D) | 143 DCE-MR cases (M: 77, B: 66) | [ |
| CNN (ResNet50) fine-tuned | AUC = 0.97–0.99 | Training: 33 patients with 153 lesions (M: 91, B: 62) | [ |
| Cross-modal DL (mammography and MR), trained from scratch | Acc = 94 | 123 DCE-MR + T1W 282 mammography images | [ |
| Dense convolutional LSTM | Acc = 0.847 | 72 lesions (M: 27, B: 45) DCE-MRI and DWI-MRI | [ |
| DenseNet | AUC = 0.811 | 576 lesions (M: 368, B: 149, FU: 59) | [ |
| CNN (AlexNet) fine-tuned from ImageNet on the second postcontrast frame, LSTM model for final prediction | Acc = 76 | 42 DCE-MR images, 67 lesions (M: 42, B: 25) 10-fold cross-validation | [ |
| CNN (ResNet34) fine-tuned best three inputs out of 85 combinations | AUC = 0.88 (95% confidence interval: 0.86–0.89) | 447 patients, 1294 lesions (M: 787, B: 507) multiparametric DCE-MR + T2W | [ |
| MIP + off-the-shelf CNN (VGG) + SVM | AUC = 0.88 ± 0.01 | 690 DCE-MR cases (M: 478, B: 212) | [ |
| Multiscale 3D CNN (trained from scratch) inputs: five timepoints T1W DCE-MR and T2W | AUC = 0.89 ± 0.01 | 408 patients (M: 305, B: 103) multiparametric | [ |
| Off-the-shelf CNN (VGG) + SVM target: different molecular subtypes | AUC = 0.65 (pretrained) | 270 DCE-MR images (90 luminal A, 180 other 3 subtypes) | [ |
| 3TP-CNN pretrained on ImageNet | Acc = 74 | 39 lesions (M: 36, B: 22) | [ |
| Three-channel (pre- and postcontrast) CNN (VGG) fine-tuned for classification | AUC = 0.88 | 703 DCE-MR dataset (M: 482, B: 221) | [ |
| 3D ResNet trained from scratch with ad hoc embedding loss weakly supervised localization with feature correlation attention map | Acc = 85.5 | 1715 subjects (M: 1137, B: 578) | [ |
For each study, we report the number of histologically verified benign (B) and malignant (M) lesions or cases; benign lesions without biopsy with at least 12-month follow-up (FU) are also indicated. Histology is used as ground truth in all studies.
Figure 5Two-step transfer learning approach for leveraging temporal information in pretrained CNN. In the first approach (a), the CNN is fine-tuned on pseudocolor ROIs, formed by the precontrast and first and second postcontrast frames, mimicking the three channels of an RGB images. In the second step (b), image features extracted from the trained CNN at each DCE timepoint are used to train an LSTM network, which learns to distinguish contrast enhancement patterns. Reprinted from [58].
Figure 6Flow diagram of sentinel lymph node prediction. Reprinted with permission from [113].