| Literature DB >> 34094903 |
Haimei Li1, Bing Liu2,3, Yongtao Zhang4, Chao Fu5, Xiaowei Han6, Lei Du2, Wenwen Gao2, Yue Chen2, Xiuxiu Liu2, Yige Wang2, Tianfu Wang4, Guolin Ma2, Baiying Lei4.
Abstract
Automatic segmentation of gastric tumor not only provides image-guided clinical diagnosis but also assists radiologists to read images and improve the diagnostic accuracy. However, due to the inhomogeneous intensity distribution of gastric tumors in CT scans, the ambiguous/missing boundaries, and the highly variable shapes of gastric tumors, it is quite challenging to develop an automatic solution. This study designs a novel 3D improved feature pyramidal network (3D IFPN) to automatically segment gastric tumors in computed tomography (CT) images. To meet the challenges of this extremely difficult task, the proposed 3D IFPN makes full use of the complementary information within the low and high layers of deep convolutional neural networks, which is equipped with three types of feature enhancement modules: 3D adaptive spatial feature fusion (ASFF) module, single-level feature refinement (SLFR) module, and multi-level feature refinement (MLFR) module. The 3D ASFF module adaptively suppresses the feature inconsistency in different levels and hence obtains the multi-level features with high feature invariance. Then, the SLFR module combines the adaptive features and previous multi-level features at each level to generate the multi-level refined features by skip connection and attention mechanism. The MLFR module adaptively recalibrates the channel-wise and spatial-wise responses by adding the attention operation, which improves the prediction capability of the network. Furthermore, a stage-wise deep supervision (SDS) mechanism and a hybrid loss function are also embedded to enhance the feature learning ability of the network. CT volumes dataset collected in three Chinese medical centers was used to evaluate the segmentation performance of the proposed 3D IFPN model. Experimental results indicate that our method outperforms state-of-the-art segmentation networks in gastric tumor segmentation. Moreover, to explore the generalization for other segmentation tasks, we also extend the proposed network to liver tumor segmentation in CT images of the MICCAI 2017 Liver Tumor Segmentation Challenge.Entities:
Keywords: CT volumes; adaptive spatial feature fusion; feature pyramidal network; feature refinement; gastric tumor segmentation
Year: 2021 PMID: 34094903 PMCID: PMC8173118 DOI: 10.3389/fonc.2021.618496
Source DB: PubMed Journal: Front Oncol ISSN: 2234-943X Impact factor: 6.244
Figure 1The flowchart of our gastric tumor segmentation network equipped with multi-type feature enhancement modules (3D ASFF, 3D adaptive spatial feature fusion; SLFR, Single-level feature refinement; MLFR, Multi-level feature refinement).
Figure 2The schematic illustration of the SLFR module and MLFR module.
Automatic segmentation results of different methods.
| Method | Dice (%) | JI (%) | Pre (%) | Recall (%) | ASD (voxel) | 95HD (voxel) |
|---|---|---|---|---|---|---|
| 3D U-Net ( | 59.4 ± 4.8 | 42.4 ± 4.8 | 64.0 ± 6.1 | 55.9 ± 6.4 | 15.3 ± 14.2 | 31.4 ± 11.1 |
| nnU-Net ( | 60.2 ± 3.5 | 43.2 ± 4.1 | 60.2 ± 6.6 |
| 18.3 ± 19.1 | 35.7 ± 22.2 |
| DAF3D ( | 60.8 ± 4.2 | 43.8 ± 4.2 | 64.1 ± 3.8 | 58.3 ± 7.0 | 14.7 ± 12.5 | 29.1 ± 10.6 |
| 3D FPN ( | 59.3 ± 3.6 | 42.2 ± 3.7 | 64.9 ± 8.9 | 56.2 ± 8.8 | 17.2 ± 15.7 | 34.6 ± 15.2 |
|
|
|
|
| 61.7 ± 5.4 |
|
|
Mean ± SD, with best results highlighted in bold.
Figure 3Comparison results of different automatic segmentation methods and backbones.
Figure 42D visual comparisons of segmented slices from 3D CT volumes. Up row, CT slicers with red boxes to indicate the tumor areas; Down row, ground truth (red) delineated by experienced radiologists and corresponding segmented tumor contours using 3D U-Net (10) (blue), nnU-Net (32) (green), DAF3D (14) (cyan), 3D FPN (12) (yellow) and our method (purple).
Figure 53D visualization of the automatic segmentation performance. Rows denote segmentation outcomes on four different CT volumes respectively. Columns demonstrate the visualized comparisons between the segmented surface (blue gridlines) and ground truth (red volumes) under five different automatic segmentation methods: (A) 3D U-Net (10), (B) nnU-Net (32), (C) DAF3D (14), (D) 3D FPN (12), and (E) our method, respectively.
Figure 63D visualization of the surface distance (in voxel) between segmented surface and ground truth, with color bars relating to various surface distances. Rows denote segmentation outcomes on four different CT volumes respectively. Columns demonstrate the segmented surfaces generated by (A) 3D U-Net (10), (B) nnU-Net (32), (C) DAF3D (14), (D) 3D FPN (12), and (E) our method, respectively.
Automatic segmentation results of ablation analyses.
| Method | Dice (%) | JI (%) | Pre (%) | Recall (%) | ASD (voxel) | 95HD (voxel) |
|---|---|---|---|---|---|---|
| Baseline + M1 | 60.0 ± 2.2 | 42.4 ± 2.6 | 64.2 ± 4.4 | 55.9 ± 2.2 | 18.4 ± 6.2 | 29.7 ± 3.5 |
| Baseline + M1 + M2 | 61.2 ± 3.3 | 44.2 ± 3.3 | 64.1 ± 3.6 | 58.9 ± 5.9 | 13.7 ± 13.5 | 27.1 ± 13.8 |
| Baseline + M1 + M2 + ASPP ( | 61.2 ± 3.3 | 44.2 ± 3.4 | 62.8 ± 2.5 | 60.4 ± 8.2 |
|
|
| Baseline + M1 + M2 + M3 | 61.6 ± 2.8 | 44.3 ± 2.8 | 65.0 ± 4.7 | 60.0 ± 6.7 | 17.5 ± 10.0 | 32.4 ± 10.2 |
| Baseline + M1 + M2 + M3 + DSN ( | 61.9 ± 2.7 | 44.9 ± 3.2 | 64.3 ± 3.4 | 59.4 ± 4.6 | 14.1 ± 7.1 | 27.7 ± 9.5 |
|
|
|
|
|
| 14.2 ± 8.9 | 28.2 ± 9.9 |
Mean ± SD, with best results highlighted in bold. Baseline: 3D SE-ResNeXt (18), M1: 3D ASFF module, M2: SLFR module, M3: MLFR module.
Automatic segmentation results based on different backbones.
| Backbone | Dice (%) | JI (%) | Pre (%) | Recall (%) | ASD (voxel) | 95HD (voxel) |
|---|---|---|---|---|---|---|
| ResNet ( | 61.4 ± 3.2 | 44.3 ± 2.9 | 63.2 ± 3.0 | 59.5 ± 4.8 | 17.8 ± 10.2 | 33.0 ± 11.3 |
| ResNeXt ( | 62.1 ± 2.8 | 45.1 ± 3.1 | 65.6 ± 3.5 | 61.0 ± 4.3 | 15.5 ± 9.5 | 30.6 ± 8.7 |
| SE-ResNeXt (Ours) |
|
|
|
|
|
|
Mean ± SD, with best results highlighted in bold.
Automatic segmentation results of different methods in LiTS challenge.
| Method | Liver Segmentation | Tumor Segmentation | ||
|---|---|---|---|---|
| Dice (%) | ASSD (voxel) | Dice (%) | ASSD (voxel) | |
| IeHealth | 96.1 | 1.13 | 70.2 | 1.19 |
| H-DenseNet ( | 96.1 | 1.69 | 72.2 | 1.07 |
| 3D DenseUNet ( | 93.6 | – | 59.4 | – |
| 3D AH-Net ( | 96.3 | 1.10 | 65.7 | 1.15 |
| Med3D ( | 94.6 | 1.90 | – | – |
| V-Net ( | 93.9 | 2.20 | – | – |
| Ours | 92.2 | 6.65 | 65.5 | 1.14 |
Mean ± SD, -represents the measurement was not evaluated.