| Literature DB >> 35634040 |
Yinan Lu1, Yan Zhao1, Xing Chen2, Xiaoxin Guo1.
Abstract
Medical multiobjective image segmentation aims to group pixels to form multiple regions based on the different properties of the medical images. Segmenting the 3D cardiovascular magnetic resonance (CMR) images is still a challenging task owing to several reasons, including individual differences in heart shapes, varying signal intensities, and differences in data signal-to-noise ratios. This paper proposes a novel and efficient U-Net-based 3D sparse convolutional network named SparseVoxNet. In this network, there are direct connections between any two layers with the same feature-map size, and the number of connections is reduced. Therefore, the SparseVoxNet can effectively cope with the optimization problem of gradients vanishing when training a 3D deep neural network model on small sample data by significantly decreasing the network depth, and achieveing better feature representation using a spatial self-attention mechanism finally. The proposed method in this paper has been thoroughly evaluated on the HVSMR 2016 dataset. Compared with other methods, the method achieves better performance.Entities:
Mesh:
Year: 2022 PMID: 35634040 PMCID: PMC9142300 DOI: 10.1155/2022/4103524
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1An overview of the proposed SparseVoxNet, with intermediate feature volumes. The light blue and dark blue areas of the slice represent the blood pool and myocardium. The dark blue and black areas belong to the background. The blue, yellow, and dark purple of segmented result represent the myocardium, blood pool, and background, respectively. There are two sparse blocks in this network. The black dotted line at the bottom right represents a skip connection.
Our 3D convolutional model.
| Input image | Output | Layer (type) | Stride | Kernel | Parameters | ||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| 64 | 64 | 64 | 1 | 32 | 32 | 32 | 16 | Conv_1 (convolution) | 2 | 3 | 448 |
| 32 | 32 | 32 | 16 | 16 | 16 | 16 | 16 | Conv_2 (convolution) | 2 | 3 | 6928 |
| 16 | 16 | 16 | 16 | 16 | 16 | 16 | 16 | Spatial attention | 2 | 1 | 816 |
| 16 | 16 | 16 | 16 | 16 | 16 | 16 | 100 | Sparse Block_1 (sparse block) | 1 | 3 | 43300 |
| 16 | 16 | 16 | 100 | 16 | 16 | 16 | 100 | Conv_3 (convolution) | 1 | 1 | 10100 |
| 16 | 16 | 16 | 100 | 16 | 16 | 16 | 184 | Sparse Block_2 (sparse block) | 1 | 3 | 496984 |
| 16 | 16 | 16 | 184 | 16 | 16 | 16 | 64 | Conv_4 (convolution) | 1 | 1 | 11840 |
| 16 | 16 | 16 | 64 | 32 | 32 | 32 | 64 | Deconv_1 (deconvolution) | 2 | 4 | 262208 |
| 32 | 32 | 32 | 64 | 64 | 64 | 64 | 64 | Deconv_2 (deconvolution) | 2 | 4 | 262208 |
| 16 | 16 | 16 | 100 | 64 | 64 | 64 | 64 | Skip connection | 1 | 1 | 6464 |
Results of ablation study. Bold results are the best ones.
| Myocardium | Blood pool | |||||
|---|---|---|---|---|---|---|
| Method | Dice (%) | ADB | Haus. | Dice (%) | ADB | Haus. |
| SparseVoxNet-D |
| 0.922 | 5.385 |
| 1.073 | 7.736 |
| SparseVoxNet-S | 80.7 |
|
| 91.4 |
|
|
| DenseVoxNet | 79.2 | 0.943 | 7.175 | 89.48 | 0.955 | 9.608 |
Figure 2Segmentation results on three training images.
Figure 3Segmentation results on three testing images: The three slices in the first line come from different patients. Images in the second line are the results of automatic segmentation by the method proposed in this paper.
Figure 4Comparison of experimental results between the improved method and other methods.
Comparison of experimental results between the improved method and the 3D methods. Bold results are the best ones.
| Myocardium | Blood pool | |||||
|---|---|---|---|---|---|---|
| Method | Dice (%) | ADB | Haus. | Dice (%) | ADB | Haus. |
| 3D U-Net [ | 69.4 | 2.596 | 12.796 | 79.4 | 2.550 | 14.634 |
| V-Net [ | 70.3 | 2.367 | 10.624 | 81.9 | 2.435 | 12.539 |
| VoxResNet [ | 77.4 | 2.041 | 13.199 | 86.7 | 2.157 | 19.723 |
| DenseVoxNet [ | 79.2 | 0.943 | 7.175 | 89.48 | 0.955 | 9.608 |
| Ours |
|
|
|
|
|
|