| Literature DB >> 35986015 |
Mohammed Yusuf Ansari1, Yin Yang2, Shidin Balakrishnan1, Julien Abinahed1, Abdulla Al-Ansari1, Mohamed Warfa3, Omran Almokdad1, Ali Barah1, Ahmed Omer1, Ajay Vikram Singh4, Pramod Kumar Meher5, Jolly Bhadra6, Osama Halabi6, Mohammad Farid Azampour7, Nassir Navab7, Thomas Wendler7, Sarada Prasad Dakua8.
Abstract
Segmentation of abdominal Computed Tomography (CT) scan is essential for analyzing, diagnosing, and treating visceral organ diseases (e.g., hepatocellular carcinoma). This paper proposes a novel neural network (Res-PAC-UNet) that employs a fixed-width residual UNet backbone and Pyramid Atrous Convolutions, providing a low disk utilization method for precise liver CT segmentation. The proposed network is trained on medical segmentation decathlon dataset using a modified surface loss function. Additionally, we evaluate its quantitative and qualitative performance; the Res16-PAC-UNet achieves a Dice coefficient of 0.950 ± 0.019 with less than half a million parameters. Alternatively, the Res32-PAC-UNet obtains a Dice coefficient of 0.958 ± 0.015 with an acceptable parameter count of approximately 1.2 million.Entities:
Mesh:
Year: 2022 PMID: 35986015 PMCID: PMC9391485 DOI: 10.1038/s41598-022-16828-6
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.996
Figure 1Lightweight Res32-PAC-UNet architecture for high accuracy liver CT segmentation.
Figure 2(a) Residual block employed in the backbone for improving information and gradient flow. (b) PAC module for capturing multi-scale volumetric features at different levels of the encoder.
Figure 3Proposed deep learning framework for training and inference of lightweight liver CT segmentation models.
Figure 4Evolution of DC during the first 50 epochs of training on the test set: (A) Res32-PAC-UNet trained with three different loss functions. (B) Proposed models trained with modified surface loss function.
Segmentation performance of Res32-PAC-UNet using different loss functions, indicating maximal performance with modified surface loss function.
| Loss function | DC | IoU | Sensitivity | Specificity | SVD | VOE |
|---|---|---|---|---|---|---|
| Focal loss | 0.898 (0.024) | 0.815 (0.038) | 0.95 | 0.102 (0.024) | 0.185 (0.038) | |
| Binary cross entropy | 0.949 (0.016) | 0.903 (0.028) | 0.965 (0.028) | 0.997 | 0.051 (0.016) | 0.097 (0.028) |
| Modified surface loss | 0.997 |
Significant values are in [bold].
Segmentation performance, disk utilization, and inference time of the proposed models with/without PAC module and related work, trained using modified surface loss.
| Model name | DC | IoU | Sensitivity | Specificity | SVD | VOE | Parameter count (Model size in MB) | Inference time (sec) |
|---|---|---|---|---|---|---|---|---|
| UNet (2016) | 0.919 (0.188) | 0.88 (0.182) | 0.922 (0.19) | 0.997 (0.002) | 0.081 (0.188) | 0.12 (0.182) | 22,575,329 (271) | 0.503 |
| Tuned-UNet | 0.955 | 0.914 | 0.959 (0.026) | 0.997 | 0.045 | 0.086 | 5,644,913 (68) | 0.266 |
| Multi-Res-UNet (2020) | 0.917 (0.025) | 0.848 (0.042) | 0.939 (0.036) | 0.993 (0.003) | 0.083 (0.025) | 0.152 (0.042) | 4,608,478 (55.8) | 0.474 |
| TMD-UNet (2021) | 0.923 (0.044) | 0.859(0.071) | 0.928 (0.071) | 0.995 (0.004) | 0.077 (0.044) | 0.141(0.071) | 9,109,969 (110) | 4.57 |
| DC-UNet (2021) | 0.95 (0.014) | 0.905(0.025) | 0.959 (0.026) | 0.996(0.002) | 0.05(0.014) | 0.095(0.025) | 7,065,285 (85.3) | 0.585 |
| Res-UNet++ (2019) | 0.956 | 0.916 (0.026) | 0.955 (0.028) | 0.997 | 0.044 | 0.084 (0.026) | 11,786,089 (142) | 2.44 |
| Thin16-PAC-UNet | 0.946 (0.017) | 0.898 (0.03) | 0.946 (0.028) | 0.997 (0.002) | 0.054 (0.017) | 0.102 (0.03) | 468,737 (5.89) | 0.298 |
| Thin32-PAC-UNet | 0.95 (0.015) | 0.905 (0.026) | 0.957 | 0.996 (0.002) | 0.05 (0.015) | 0.095 (0.026) | 1,202,209 (14.81) | 0.497 |
| Res16-UNet | 0.931 (0.04) | 0.873 (0.063) | 0.933 (0.037) | 0.995 (0.011) | 0.069 (0.04) | 0.127 (0.063) |
|
|
| Res32-UNet | 0.954 (0.014) | 0.912 (0.025) | 0.952 (0.026) | 0.997 | 0.046 | 0.088 | 627,521 (7.82) | 0.442 |
| Res16-PAC-UNet | 0.95 (0.019) | 0.905 (0.033) | 0.942 (0.029) | 0.05 (0.019) | 0.095 (0.033) | 478,849 (6.15) | 0.320 | |
| Res32-PAC-UNet | 0.997 | 1,227,041(15.1) | 0.525 |
Significant values are in [bold].
Figure 5Qualitative comparison of the different segmentation masks generated by the proposed neural networks. The red bounding oval marks the presence of artifacts. The predicted segmentation masks (yellow) are overlaid on the ground truth (red) to highlight region overlap.