| Literature DB >> 34209154 |
Justin Lo1,2, Saiee Nithiyanantham1,2, Jillian Cardinell1,2, Dylan Young1,2, Sherwin Cho2, Abirami Kirubarajan3, Matthias W Wagner4, Roxana Azma4, Steven Miller5, Mike Seed6, Birgit Ertl-Wagner4, Dafna Sussman1,2,7,8.
Abstract
Segmentation of the fetus from 2-dimensional (2D) magnetic resonance imaging (MRI) can aid radiologists with clinical decision making for disease diagnosis. Machine learning can facilitate this process of automatic segmentation, making diagnosis more accurate and user independent. We propose a deep learning (DL) framework for 2D fetal MRI segmentation using a Cross Attention Squeeze Excitation Network (CASE-Net) for research and clinical applications. CASE-Net is an end-to-end segmentation architecture with relevant modules that are evidence based. The goal of CASE-Net is to emphasize localization of contextual information that is relevant in biomedical segmentation, by combining attention mechanisms with squeeze-and-excitation (SE) blocks. This is a retrospective study with 34 patients. Our experiments have shown that our proposed CASE-Net achieved the highest segmentation Dice score of 87.36%, outperforming other competitive segmentation architectures.Entities:
Keywords: attention mechanisms; automatic segmentation; convolutional neural networks; deep learning; fetal MRI; squeeze-and-excitation
Mesh:
Year: 2021 PMID: 34209154 PMCID: PMC8272176 DOI: 10.3390/s21134490
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1The given MRI dataset is applied through an automatic segmentation algorithm for clinical diagnosis.
Figure 2An overview of the proposed CASE-Net architecture.
Figure 3Block diagram of the SE block.
Figure 4Block diagram of the Attention Mechanism.
MSE of the Dataset Compared with the Output Produced by the Algorithm.
| Dataset 1 | Dataset 2 | Dataset 3 | Dataset 4 | Dataset 5 | Dataset 6 | |
|---|---|---|---|---|---|---|
| Person 1 | 495.54 | 366.36 | 1101.5 | 612.45 | 444.3 | 456.48 |
| Person 2 | 465.54 | 309.99 | 1158.5 | 522.39 | 350.33 | 370.46 |
Aggregated Mean Testing Results with Standard Deviation.
| Model | Loss | DSC | Recall | Precision |
|---|---|---|---|---|
| CASE-Net | 0.005 | 87.36 ± 0.83 | 91.79 ± 0.44 | 95.54 ± 0.13 |
| UNet | 0.069 | 85.17 ± 0.41 | 89.24 ± 0.26 | 94.55 ± 0.17 |
| USE-Net | 0.04 | 86.07 ± 0.41 | 86.61 ± 2.49 | 96.74 ± 0.27 |
| ATN-Net | 0.131 | 82.27 ± 0.47 | 88.69 ± 1.25 | 96.51 ± 0.27 |
| Link-Net | 0.096 | 82.72 ± 0.61 | 80.39 ± 4.96 | 95.86 ± 0.27 |
CASE-Net Ablation Study with Standard Deviation.
| Method | Loss | DSC | Recall | Precision |
|---|---|---|---|---|
| CASE-Net | 0.005 | 87.36 ± 0.83 | 91.79 ± 0.44 | 95.54 ± 0.13 |
| W/3 × 3 Kernel | 0.052 | 83.50 ± 0.35 | 86.51 ± 2.35 | 96.47 ± 0.24 |
| W/O ATN | 0.018 | 84.35 ± 0.13 | 89.06 ± 1.78 | 96.26 ± 0.20 |
| W/O SE | 0.11 | 81.19 ± 1.20 | 80.95 ± 4.20 | 96.69 ± 0.40 |
Figure 5Boxplot representation of the DSC on various segmentation architectures.
Figure 6Predicted segmentation outputs of various architectures. (A): Label, (B): CASE-Net, (C): U-Net, (D): USE-Net, (E): ATN-Net, (F): Link-Net.
Figure 7Pixel-to-pixel comparison of the ground truth mask against the predicted segmentations of various architectures. (A): Label, (B): CASE-Net, (C): U-Net, (D): USE-Net, (E): ATN-Net, (F): Link-Net. TP is given by the color white, TN is given by the color grey, FP is given by the color blue, and FN is given by the color red.