| Literature DB >> 36068814 |
Ge Jin1, Chuancai Liu1,2, Xu Chen1.
Abstract
Since the outbreak of Coronavirus Disease 2019 (COVID-19) in 2020, it has significantly affected the global health system. The use of deep learning technology to automatically segment pneumonia lesions from Computed Tomography (CT) images can greatly reduce the workload of physicians and expand traditional diagnostic methods. However, there are still some challenges to tackle the task, including obtaining high-quality annotations and subtle differences between classes. In the present study, a novel deep neural network based on Resnet architecture is proposed to automatically segment infected areas from CT images. To reduce the annotation cost, a Vector Quantized Variational AutoEncoder (VQ-VAE) branch is added to reconstruct the input images for purpose of regularizing the shared decoder and the latent maps of the VQ-VAE are utilized to further improve the feature representation. Moreover, a novel proportions loss is presented for mitigating class imbalance and enhance the generalization ability of the model. In addition, a semi-supervised mechanism based on adversarial learning to the network has been proposed, which can utilize the information of the trusted region in unlabeled images to further regularize the network. Extensive experiments on the COVID-SemiSeg are performed to verify the superiority of the proposed method, and the results are in line with expectations.Entities:
Keywords: Adversarial network; COVID-19; Infection segmentation; Proportions loss; Semi-supervised learning; VQ-VAE
Year: 2022 PMID: 36068814 PMCID: PMC9436790 DOI: 10.1016/j.ins.2022.08.059
Source DB: PubMed Journal: Inf Sci (N Y) ISSN: 0020-0255 Impact factor: 8.233
Fig. 1Overview of the proposed model.
Fig. 2The visual comparison between the proposed method and various losses. The color masks denote the COVID-19 infected regions in CT axial slice, where the red and green denote the GGO and consolidation respectively. The images in the left-most column are the ground-truth. The examples show that our method can achieve the best performance. The proposed method is susceptible to the segmentation regions of both classes. However, for the first example, CE basically did not identify the consolidation region. Moreover, the combinations of the CE and Dice also lost some details.
Quantitative results of ground-glass opacities and consolidation on the COVID-SemiSeg dataset. The best and the second results are highlighted in red and blue.
| Ground-Glass Opacity | Consolidation | Average | ||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Methods | Dice | Sen. | Spec. | MAE | Dice | Sen. | Spec. | MAE | Dice | Sen. | Spec. | MAE | ||||||
| FCN | 0.480 | 0.450 | 0.910 | 0.582 | 0.754 | 0.101 | 0.283 | 0.268 | 0.714 | 0.554 | 0.561 | 0.051 | 0.382 | 0.359 | 0.812 | 0.568 | 0.658 | 0.076 |
| Deeplabv3+ | 0.462 | 0.591 | 0.912 | 0.551 | 0.690 | 0.079 | 0.201 | 0.289 | 0.685 | 0.598 | 0.610 | 0.049 | 0.332 | 0.440 | 0.799 | 0.565 | 0.650 | 0.064 |
| U-Net | 0.473 | 0.530 | 0.944 | 0.577 | 0.743 | 0.098 | 0.274 | 0.302 | 0.673 | 0.640 | 0.805 | 0.048 | 0.374 | 0.416 | 0.809 | 0.609 | 0.774 | 0.073 |
| U-Net++ | 0.488 | 0.542 | 0.952 | 0.580 | 0.787 | 0.080 | 0.280 | 0.311 | 0.745 | 0.615 | 0.820 | 0.046 | 0.384 | 0.427 | 0.849 | 0.598 | 0.740 | 0.063 |
| Attention UNet | 0.491 | 0.539 | 0.961 | 0.576 | 0.792 | 0.082 | 0.279 | 0.320 | 0.752 | 0.597 | 0.811 | 0.049 | 0.385 | 0.430 | 0.857 | 0.587 | 0.802 | 0.066 |
| R2U-Net | 0.410 | 0.481 | 0.875 | 0.554 | 0.716 | 0.097 | 0.193 | 0.284 | 0.693 | 0.495 | 0.554 | 0.049 | 0.302 | 0.383 | 0.784 | 0.525 | 0.635 | 0.070 |
| PSPNet | 0.500 | 0.445 | 0.945 | 0.594 | 0.820 | 0.081 | 0.241 | 0.279 | 0.687 | 0.655 | 0.794 | 0.044 | 0.371 | 0.343 | 0.816 | 0.625 | 0.807 | 0.063 |
| MANet | 0.537 | 0.519 | 0.969 | 0.633 | 0.880 | 0.078 | 0.292 | 0.285 | 0.743 | 0.651 | 0.797 | 0.044 | 0.415 | 0.402 | 0.849 | 0.642 | 0.839 | 0.061 |
| 0.587 | 0.606 | 0.963 | 0.644 | 0.877 | 0.073 | 0.323 | 0.333 | 0.759 | 0.678 | 0.855 | 0.039 | 0.455 | 0.470 | 0.861 | 0.671 | 0.866 | 0.056 | |
Fig. 3Visual comparison of lung infection segmentation results, where the red and green regions indicate the GGO and consolidation, respectively.
Ablation study of the proposed methods. stands for VQ-VAE module, denotes semi-supervised adversarial learning. The best two results are shown in red and blue fonts.
| Ground-Glass Opacity | Consolidation | |||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Methods | Dice | Sen. | Spec. | MAE | Dice | Sen. | Spec. | MAE | ||||
| Backbone | 0.473 | 0.575 | 0.891 | 0.591 | 0.574 | 0.140 | 0.187 | 0.261 | 0.613 | 0.573 | 0.759 | 0.049 |
| Backbone+ | 0.520 | 0.579 | 0.912 | 0.599 | 0.659 | 0.136 | 0.213 | 0.279 | 0.628 | 0.621 | 0.775 | 0.048 |
| Backbone+ | 0.514 | 0.587 | 0.935 | 0.510 | 0.731 | 0.105 | 0.251 | 0.284 | 0.662 | 0.588 | 0.802 | 0.045 |
| Backbone+ | 0.539 | 0.599 | 0.944 | 0.625 | 0.854 | 0.128 | 0.249 | 0.278 | 0.690 | 0.597 | 0.783 | 0.043 |
| Backbone+ | 0.557 | 0.590 | 0.938 | 0.618 | 0.810 | 0.095 | 0.279 | 0.301 | 0.702 | 0.630 | 0.791 | 0.043 |
| Backbone+ | 0.571 | 0.609 | 0.956 | 0.641 | 0.838 | 0.091 | 0.294 | 0.315 | 0.713 | 0.628 | 0.811 | 0.041 |
| 0.587 | 0.606 | 0.963 | 0.644 | 0.877 | 0.073 | 0.323 | 0.333 | 0.759 | 0.678 | 0.855 | 0.039 | |
Fig. 4Visualization results of the reconstructed images and confidence maps.