| Literature DB >> 34659447 |
Jinlin Ma1,2, Yuanyuan Deng1, Ziping Ma3, Kaiji Mao1, Yong Chen4.
Abstract
Accurate segmentation of liver images is an essential step in liver disease diagnosis, treatment planning, and prognosis. In recent years, although liver segmentation methods based on 2D convolutional neural networks have achieved good results, there is still a lack of interlayer information that causes severe loss of segmentation accuracy to a certain extent. Meanwhile, making the best of high-level and low-level features more effectively in a 2D segmentation network is a challenging problem. Therefore, we designed and implemented a 2.5-dimensional convolutional neural network, VNet_WGAN, to improve the accuracy of liver segmentation. First, we chose three adjacent layers of a liver model as the input of our network and adopted two convolution kernels in series connection, which can integrate cross-sectional spatial information and interlayer information of liver models. Second, a chain residual pooling module is added to fuse multilevel feature information to optimize the skip connection. Finally, the boundary loss function in the generator is employed to compensate for the lack of marginal pixel accuracy in the Dice loss function. The effectiveness of the proposed method is verified on two datasets, LiTS and CHAOS. The Dice coefficients are 92% and 90%, respectively, which are better than those of the compared segmentation networks. In addition, the experimental results also show that the proposed method can reduce computational consumption while retaining higher segmentation accuracy, which is significant for liver segmentation in practice and provides a favorable reference for clinicians in liver segmentation.Entities:
Mesh:
Year: 2021 PMID: 34659447 PMCID: PMC8519672 DOI: 10.1155/2021/5536903
Source DB: PubMed Journal: Comput Math Methods Med ISSN: 1748-670X Impact factor: 2.238
Figure 1Basic structure of GAN.
Figure 2Structure diagram of VNet_WGAN.
Figure 3Improved convolution module.
Figure 4Chain residual pooling module.
Figure 5Structure diagram of liver segmentation generator.
Network parameter.
| Network layer | Number of feature maps | Feature map size/pixels | Number of participants/each |
|---|---|---|---|
| Convolutional layer 1 | 32 | 128 × 128 × 128 | 896 |
| Convolutional layer 2 | 32 | 64 × 64 × 64 | 27680 |
| Convolutional layer 3 | 64 | 32 × 32 × 32 | 110656 |
| Convolutional layer 4 | 128 | 16 × 16 × 16 | 442496 |
| Convolutional layer 5 | 256 | 8 × 8 × 8 | 1769728 |
| Convolutional layer 6 | 256 | 4 × 4 × 4 | 1769728 |
| Deconvolution 1 | 512 | 8 × 8 × 8 | 3539456 |
| Convolutional layer 7 | 256 | 8 × 8 × 8 | 5308672 |
| Deconvolution 2 | 256 | 8 × 8 × 8 | 1769728 |
| Convolutional layer 8 | 128 | 16 × 16 × 16 | 1327232 |
| Deconvolution 3 | 128 | 32 × 32 × 32 | 442496 |
| Convolutional layer 9 | 64 | 32 × 32 × 32 | 110656 |
| Deconvolution 4 | 64 | 64 × 64 × 64 | 82976 |
| Convolutional layer 10 | 32 | 64 × 64 × 64 | 82976 |
| Deconvolution 5 | 64 | 64 × 64 × 64 | 82976 |
| Convolutional layer 11 | 32 | 128 × 128 × 128 | 27680 |
| Convolutional layer 12 | 32 | 128 × 128 × 128 | 55328 |
| Fully connected layer | 1 | 128 × 128 × 128 | 33 |
| Total | - | - | 22482273 |
Note: “-” means no data.
Figure 6Structure of liver segmentation discriminant.
Figure 7Parameter relationship.
Figure 8Example of liver tumor images in the LiTS dataset.
Figure 9Example of liver images in the CHAOS dataset.
Figure 10Concepts related to evaluation indicators.
The experimental results of different methods were compared.
| Comparison method | Dataset | Accuracy | Dice |
|---|---|---|---|
| 3D UNet | LiTS | 0.98 | 0.91 |
| 3D UNet | CHAOS | 0.96 | 0.89 |
| Improved 2.5D VNet | LiTS | 0.96 | 0.89 |
| 3D VNet | LiTS | 0.95 | 0.90 |
| 2.5D UNet | LiTS | — | 0.77 |
| Attention [ | LiTS | — | 0.76 |
| CycleGAN [ | LiTS | 0.86 | 0.89 |
| DenseNet | LiTS | — | 0.91 |
| Ours (VNet_WGAN) | LiTS | 0.94 | 0.92 |
| Ours (VNet_WGAN) | CHAOS | 0.94 | 0.90 |
Note: “-” means no data.
Comparison of segmentation effects of different loss functions in the 2.5D VNet model.
| Loss function | Accuracy | Dice |
|---|---|---|
| IoU loss | 0.96 | 0.80 |
| Cross-entropy loss | 0.95 | 0.84 |
| Dice loss | 0.96 | 0.88 |
| Boundary loss |
| 0.83 |
| Dice loss+boundary loss (ours) | 0.96 |
|
Note: the bold font is the optimal value for each column.
Comparison of segmentation effects of different loss functions in VNet and WGAN fusion models (ours).
| Loss function | Accuracy | Dice |
|---|---|---|
| IoU loss | 0.93 | 0.86 |
| Cross-entropy loss |
| 0.89 |
| Dice loss | 0.94 | 0.91 |
| Boundary loss | 0.92 | 0.88 |
| Dice loss+boundary loss (ours) | 0.94 |
|
Note: the bold font is the optimal value for each column.
Figure 11Comparison of prediction results.