| Literature DB >> 34749094 |
Anas M Tahir1, Muhammad E H Chowdhury2, Amith Khandakar3, Tawsifur Rahman4, Yazan Qiblawey5, Uzair Khurshid6, Serkan Kiranyaz7, Nabil Ibtehaz8, M Sohel Rahman9, Somaya Al-Maadeed10, Sakib Mahmud11, Maymouna Ezeddin12, Khaled Hameed13, Tahir Hamid14.
Abstract
The immense spread of coronavirus disease 2019 (COVID-19) has left healthcare systems incapable to diagnose and test patients at the required rate. Given the effects of COVID-19 on pulmonary tissues, chest radiographic imaging has become a necessity for screening and monitoring the disease. Numerous studies have proposed Deep Learning approaches for the automatic diagnosis of COVID-19. Although these methods achieved outstanding performance in detection, they have used limited chest X-ray (CXR) repositories for evaluation, usually with a few hundred COVID-19 CXR images only. Thus, such data scarcity prevents reliable evaluation of Deep Learning models with the potential of overfitting. In addition, most studies showed no or limited capability in infection localization and severity grading of COVID-19 pneumonia. In this study, we address this urgent need by proposing a systematic and unified approach for lung segmentation and COVID-19 localization with infection quantification from CXR images. To accomplish this, we have constructed the largest benchmark dataset with 33,920 CXR images, including 11,956 COVID-19 samples, where the annotation of ground-truth lung segmentation masks is performed on CXRs by an elegant human-machine collaborative approach. An extensive set of experiments was performed using the state-of-the-art segmentation networks, U-Net, U-Net++, and Feature Pyramid Networks (FPN). The developed network, after an iterative process, reached a superior performance for lung region segmentation with Intersection over Union (IoU) of 96.11% and Dice Similarity Coefficient (DSC) of 97.99%. Furthermore, COVID-19 infections of various shapes and types were reliably localized with 83.05% IoU and 88.21% DSC. Finally, the proposed approach has achieved an outstanding COVID-19 detection performance with both sensitivity and specificity values above 99%.Entities:
Keywords: COVID-19; Chest X-ray; Convolutional Neural Networks; Deep Learning; Infection Segmentation; Lung Segmentation
Mesh:
Year: 2021 PMID: 34749094 PMCID: PMC8556687 DOI: 10.1016/j.compbiomed.2021.105002
Source DB: PubMed Journal: Comput Biol Med ISSN: 0010-4825 Impact factor: 4.589
Fig. 1Sample chest X-ray images from the COVID-QU-Ex dataset for Normal, Non-COVID, and COVID-19 classes. All images are rescaled with the same factor to illustrate the diversity of the dataset.
Fig. 2Collaborative human-machine approach to create ground-truth lung segmentation masks for COVID-QU-Ex CXR dataset. Stage Ⅰ: Three segmentation networks are trained on a repository of 704 CXR lung segmentation masks, and the best network in terms of DSC is selected for the subsequent stages. Stage Ⅱ: An iterative training is utilized to create lung masks for a subset of 3000 CXR samples from the COVID-QU-Ex dataset. Firstly, A subset of 500 samples is inferred by the CXR segmentation model and the outputs are evaluated manually as accept, reject, modify, or exclude. Next, the modified masks are added to the lung repository and the network is re-trained on the extended dataset. These steps are repeated until generating ground-truth masks for the 3000 CXR samples is completed. Stage Ⅲ: six deep segmentation networks are trained using the 3000 ground-truth masks generated in the previous stage. The trained networks are used to predict segmentation masks for the rest of the COVID-QU-Ex dataset (30,920 images). Stage Ⅳ: a final verification is performed by MDs on randomly selected 6788 CXR samples (20% of the full dataset) that well presents the diversity of the COVID-QU-Ex dataset.
Fig. 3Schematic representation of the pipeline of the proposed system. The input CXR image is fed to two ED-CNNs in parallel, to generate two binary masks: lung, and COVID-19 infection masks. Next, the generated masks are superimposed with the CXR image to localize and quantify COVID-19 infected lung regions. Finally, the generated infection mask is used to detect COVID-19 positive cases from COVID-19 negative cases.
Number of mages per class per train, validation, and test sets for each of the 5 folds used for lung segmentation, infection segmentation, and COVID-19 detection tasks.
| Dataset Name | Task | Class | # of Samples | Training Samples | Validation Samples | Test Samples | |
|---|---|---|---|---|---|---|---|
| COVID-QU-Ex dataset | Lung | COVID-19 | 11,956 | 7658 | 1903 | 2395 | |
| Non-COVID | 11,263 | 7208 | 1802 | 2253 | |||
| Normal | 10,701 | 6849 | 1712 | 2140 | |||
| COVID-QU-Ex and QaTa-Cov19 [ | Infection | COVID-19 positive | 2913 | 1864 | 466 | 583 | |
| COVID-19 negative | Non-COVID | 1457 | 932 | 233 | 292 | ||
| Normal | 1456 | 932 | 233 | 291 | |||
Performance metrics (%) for lung region and COVID-19 infected region segmentation computed over test (unseen) set with three network models and five encoder architectures. x ± y means that the achieved metric value is x with standard deviation y.
| Task | Model | Encoder | Accuracy | IoU | DSC |
|---|---|---|---|---|---|
| Lung | U-Net | ResNet18 | 99.07 ± 0.23 | 95.91 ± 0.47 | 97.88 ± 0.34 |
| ResNet50 | 99.08 ± 0.23 | 95.93 ± 0.47 | 97.89 ± 0.34 | ||
| DenseNet161 | 99.1 ± 0.22 | 96.02 ± 0.47 | 97.94 ± 0.34 | ||
| InceptionV4 | 99.07 ± 0.23 | 95.9 ± 0.47 | 97.88 ± 0.34 | ||
| U-Net ++ | ResNet18 | 99.07 ± 0.23 | 95.9 ± 0.47 | 97.88 ± 0.34 | |
| ResNet50 | 99.1 ± 0.22 | 96.04 ± 0.46 | 97.95 ± 0.34 | ||
| DenseNet161 | 99.09 ± 0.23 | 95.98 ± 0.47 | 97.92 ± 0.34 | ||
| InceptionV4 | 99.08 ± 0.23 | 95.96 ± 0.47 | 97.91 ± 0.34 | ||
| FPN | ResNet18 | 99.06 ± 0.23 | 95.86 ± 0.47 | 97.86 ± 0.34 | |
| ResNet50 | 99.07 ± 0.23 | 95.91 ± 0.47 | 97.88 ± 0.34 | ||
| DenseNet161 | 99.09 ± 0.23 | 96.01 ± 0.47 | 97.94 ± 0.34 | ||
| InceptionV4 | 99.07 ± 0.23 | 95.92 ± 0.47 | 97.89 ± 0.34 | ||
| Infection | U-Net | ||||
| ResNet50 | 97.84 ± 0.83 | 81.73 ± 2.22 | 87.02 ± 1.93 | ||
| DenseNet121 | 97.98 ± 0.81 | 82.53 ± 2.18 | 87.74 ± 1.88 | ||
| DenseNet161 | 97.86 ± 0.83 | 81.95 ± 2.21 | 87.19 ± 1.92 | ||
| InceptionV4 | 97.98 ± 0.81 | 82.03 ± 2.2 | 87.11 ± 1.92 | ||
| U-Net ++ | ResNet18 | 97.9 ± 0.82 | 82.9 ± 2.16 | 88.06 ± 1.86 | |
| ResNet50 | 97.93 ± 0.82 | 82.59 ± 2.18 | 87.78 ± 1.88 | ||
| DenseNet161 | 97.95 ± 0.81 | 81.55 ± 2.23 | 86.66 ± 1.95 | ||
| InceptionV4 | 97.9 ± 0.82 | 81.13 ± 2.25 | 86.22 ± 1.98 | ||
| FPN | ResNet18 | 97.84 ± 0.83 | 81.9 ± 2.21 | 87.25 ± 1.91 | |
| ResNet50 | 97.84 ± 0.83 | 80.83 ± 2.26 | 86.25 ± 1.98 | ||
| DenseNet121 | 97.99 ± 0.81 | 82.55 ± 2.18 | 87.71 ± 1.88 | ||
| DenseNet161 | 97.95 ± 0.81 | 81.89 ± 2.21 | 87.08 ± 1.93 | ||
Fig. 4Sample qualitative evaluation of generated lung masks by the three top-performing networks. Column 1 shows the CXR image, Column 2 shows ground truths, and the lung masks of the top three networks are shown in Columns 3–5, respectively.
Fig. 5(a) Sample qualitative evaluation of generated infection masks by the three top-performing networks. Column 1 shows the CXR image, Column 2 shows ground truths, and the lung masks of the top three networks are shown in Columns 3–5, respectively. (b) Infection localization and severity grading of COVID-19 pneumonia for a 42-year female patient on the 1st, 2nd, and 3rd days of admission using the proposed system.
COVID-19 detection performance results (%) computed over test (unseen) set with three network models, and five encoder architectures. x ± y means that the achieved metric value is x with standard deviation y.
| Model | Encoder | Accuracy | Precision | Sensitivity | F1-score | Specificity |
|---|---|---|---|---|---|---|
| U-Net | ResNet18 | 98.89 ± 0.6 | 99.14 ± 0.53 | 98.63 ± 0.67 | 98.88 ± 0.6 | 99.14 ± 0.53 |
| ResNet50 | 98.89 ± 0.6 | 98.47 ± 0.7 | 99.31 ± 0.48 | 98.89 ± 0.6 | 98.46 ± 0.71 | |
| DenseNet121 | 98.8 ± 0.62 | 97.98 ± 0.81 | 98.81 ± 0.62 | 97.94 ± 0.82 | ||
| DenseNet161 | 98.71 ± 0.65 | 97.97 ± 0.81 | 99.49 ± 0.41 | 98.72 ± 0.65 | 97.94 ± 0.82 | |
| InceptionV4 | 98.03 ± 0.8 | 98.28 ± 0.75 | 97.77 ± 0.85 | 98.02 ± 0.8 | 98.28 ± 0.75 | |
| U-Net ++ | ResNet18 | 99.23 ± 0.5 | 100 ± 0 | 98.46 ± 0.71 | 99.22 ± 0.5 | |
| ResNet50 | 99.14 ± 0.53 | 99.83 ± 0.24 | 98.46 ± 0.71 | 99.14 ± 0.53 | 99.83 ± 0.24 | |
| DenseNet121 | 99.23 ± 0.5 | 99.14 ± 0.53 | 99.22 ± 0.5 | 99.14 ± 0.53 | ||
| DenseNet161 | 98.2 ± 0.76 | 97.95 ± 0.81 | 98.46 ± 0.71 | 98.2 ± 0.76 | 97.94 ± 0.82 | |
| InceptionV4 | 98.2 ± 0.76 | 98.45 ± 0.71 | 97.94 ± 0.82 | 98.19 ± 0.77 | 98.46 ± 0.71 | |
| FPN | ResNet18 | 98.54 ± 0.69 | 97.48 ± 0.9 | 98.56 ± 0.68 | 97.43 ± 0.91 | |
| ResNet50 | 98.46 ± 0.71 | 98.46 ± 0.71 | 98.46 ± 0.71 | 98.46 ± 0.71 | 98.46 ± 0.71 | |
| DenseNet121 | 98.97 ± 0.58 | 99.65 ± 0.34 | 98.28 ± 0.75 | 98.96 ± 0.58 | 99.66 ± 0.33 | |
| DenseNet161 | 98.11 ± 0.78 | 97.3 ± 0.93 | 98.97 ± 0.58 | 98.13 ± 0.78 | 97.26 ± 0.94 | |
| InceptionV4 | 99.23 ± 0.5 | 99.31 ± 0.48 | 99.14 ± 0.53 | 99.22 ± 0.5 | 99.31 ± 0.48 |
The number of trainable parameters of the models with their inference time (ms) per CXR sample.
| Model | Encoder | Trainable parameters | Inference Time (ms) |
|---|---|---|---|
| U-Net | ResNet18 | 14.32 M | 5.78 |
| ResNet50 | 32.5 M | 10.44 | |
| DenseNet121 | 13.60 M | 22.86 | |
| DenseNet161 | 38.73 M | 29.74 | |
| InceptionV4 | 48.79 M | 26.53 | |
| U-Net ++ | ResNet18 | 15.96 M | 8.30 |
| ResNet50 | 48.97 M | 19.90 | |
| DenseNet121 | 30.06 M | 25.13 | |
| DenseNet161 | 79.04 M | 48.62 | |
| InceptionV4 | 59.35 M | 32.53 | |
| FPN | ResNet18 | 13.04 M | 5.74 |
| ResNet50 | 26.11 M | 10.34 | |
| DenseNet121 | 9.29 M | 22.68 | |
| DenseNet161 | 29.49 M | 29.62 | |
| InceptionV4 | 43.57 M | 26.08 |
Comparing the proposed work with recent literature about automatic COVID-19 diagnosis using CXR images, in terms of utilized Dataset, whether Lung and/or Infection Segmentation models are used, deployed Networks, and achieved Results.
| Ref. | Dataset (# of subjects) | Lung | Infection Seg. | Network | Results |
|---|---|---|---|---|---|
| [ | COVID19 (224) | ||||
| [ | COVID19 (358) | ||||
| [ | COVID-19 (403) | ||||
| [ | COVID-19 (423) | ||||
| [ | COVID-19 (462) | ||||
| [ | COVID-19 (500) | ||||
| [ | COVID-19 (180) | ✓ | |||
| [ | COVID-19 (423) | ✓ | |||
| [ | COVID-19 (573) | ✓ | |||
| [ | COVID-19 (2951) | ✓ | |||
| This work | COVID-19 (11,956) | ✓ | ✓ |