| Literature DB >> 35576824 |
Mei-Ling Huang1, Yu-Chieh Liao2.
Abstract
BACKGROUND AND OBJECTIVES: The traditional method of detecting COVID-19 disease mainly rely on the interpretation of computer tomography (CT) or X-ray images (X-ray) by doctors or professional researchers to identify whether it is COVID-19 disease, which is easy to cause identification mistakes. In this study, the technology of convolutional neural network is expected to be able to efficiently and accurately identify the COVID-19 disease.Entities:
Keywords: COVID-19; Computer tomography; Transfer learning; X-ray
Mesh:
Year: 2022 PMID: 35576824 PMCID: PMC9090861 DOI: 10.1016/j.compbiomed.2022.105604
Source DB: PubMed Journal: Comput Biol Med ISSN: 0010-4825 Impact factor: 6.698
Fig. 1Structure of this research.
Dataset 1: Chest X-ray image dataset.
| Category | Total number of images | Number of images selected in this study | train | validation | test |
|---|---|---|---|---|---|
| COVID-19 | 3,616 | 600 | 400 | 100 | 100 |
| Pneumonia | 5,618 | 600 | 400 | 100 | 100 |
| Normal | 11,775 | 600 | 400 | 100 | 100 |
| 21,009 | 1800 | 1,200 | 300 | 300 |
Dataset 2: Chest CT image dataset.
| Category | Total number of images | Number of images selected in this study | train | validation | test |
|---|---|---|---|---|---|
| COVID-19 | 9,003 | 600 | 400 | 100 | 100 |
| Pneumonia | 1,120 | 600 | 400 | 100 | 100 |
| Normal | 7,063 | 600 | 400 | 100 | 100 |
| 17,186 | 1800 | 1,200 | 300 | 300 |
Fig. 2Resized Chest X-ray image.
Fig. 3Resized Chest CT image.
Comparison of model parameters before and after fine-tuning.
| Before fine-tuning | Parameters | After fine-tuning | Parameters |
|---|---|---|---|
| InceptionV3 | 21,808,931 | FT-InceptionV3 | 913,299 |
| ResNet50V2 | 23,570,947 | FT-ResNet50V2 | 957,827 |
| Xception | 20,867,627 | FT-Xception | 925,083 |
| DenseNet121 | 7,040,579 | FT-DenseNet121 | 948,611 |
| MobileNetV2 | 2,261,827 | FT-MobileNetV2 | 890,851 |
| EfficientNet-B0 | 4,053,414 | FT-EfficientNet-B0 | 912,123 |
| EfficientNetV2-S | 20,181,331 | FT-EfficientNetV2-S | 863,895 |
Fig. 4Architecture of fine-tuning transfer learning models.
Fig. 5Architecture of LightEfficientNetV2.
Architecture of LightEfficientNetV2.
| Layer (type) | Filter | Kernel | Stride | Padding | Output shape | |
|---|---|---|---|---|---|---|
| Image Input | – | – | – | – | 224 × 224 × 1 | |
| Conv 1a, k 11 × 11 | 96 | 11 × 11 | 4 | 1 | 56 × 56 × 96 | |
| Batch normalization 1b | – | – | – | – | 56 × 56 × 96 | |
| Conv 2a, k 5 × 5 | 128 | 5 × 5 | 1 | 0 | 56 × 56 × 128 | |
| Batch normalization 2b | – | – | – | – | 56 × 56 × 128 | |
| Fused-MBConv3 | Conv 3a | 24 | 3 × 3 | 1 | 0 | 56 × 56 × 24 |
| Batch normalization 3b | – | – | – | – | 56 × 56 × 24 | |
| SE 3c | 24 | – | – | – | 56 × 56 × 24 | |
| 8 | – | – | – | 56 × 56 × 8 | ||
| Conv 3d | 24 | 1 × 1 | 2 | – | 28 × 28 × 24 | |
| Batch normalization 3e | – | – | – | – | 28 × 28 × 24 | |
| MBConv4 | Conv 4a | 256 | 1 × 1 | 1 | 0 | 28 × 28 × 256 |
| Batch normalization 4b | – | – | – | – | 28 × 28 × 256 | |
| SE 4c | 256 | – | – | – | 28 × 28 × 256 | |
| 8 | – | – | – | 28 × 28 × 8 | ||
| DepthwiseConv 4d | 128 | 3 × 3 | 2 | – | 14 × 14 × 128 | |
| Batch normalization 4e | – | – | – | – | 14 × 14 × 128 | |
| Conv 4f | 512 | 1 × 1 | 1 | – | 14 × 14 × 512 | |
| Batch normalization 4g | – | – | – | – | 7 × 7 × 512 | |
| Conv 5a, k 3 × 3 | 64 | 3 × 3 | 1 | 0 | 7 × 7 × 64 | |
| Batch normalization 5b | – | – | – | – | 7 × 7 × 64 | |
| Conv 6a, k 1 × 1 | 64 | 1 × 1 | 1 | 0 | 7 × 7 × 64 | |
| Batch normalization 6b | – | – | – | – | 7 × 7 × 64 | |
| Global Average Pooling | – | – | – | – | 1 × 1 × 64 | |
| Dropout | – | – | – | – | – | |
| FC | – | – | – | – | 1 × 1 × 3 | |
| Softmax | – | – | – | – | 1 × 1 × 3 | |
| Classification Output | – | – | – | – | – | |
Parameter setting.
| Parameter | Value |
|---|---|
| Batch size | 8, 16, 32 |
| Epochs | 20, 50, 100, 150, 200, 250, 300 |
| Optimizer | SGD, RMSprop, Adagrad, Adadelta, Adam, Adamax, Nadam |
| Learning rate | 0.00001, 0.0001, 0.001, 0.01, 0.1 |
| Dropout | 0.0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9 |
Confusion matrix.
| Confusion matrix | Actual situation | |||
|---|---|---|---|---|
| COVID-19 | Pneumonia | Normal | ||
| Model classification | COVID-19 | COVID-19 category that are correctly classified as COVID-19. | Pneumonia category that are wrongly classified as COVID-19. | Normal category that are wrongly classified as COVID-19. |
| Pneumonia | COVID-19 category that are wrongly classified as Pneumonia. | Pneumonia category that are correctly classified as Pneumonia. | Normal category that are wrongly classified as Pneumonia. | |
| Normal | COVID-19 category that are wrongly classified as Normal. | Pneumonia category that are wrongly classified as Normal. | Normal category that are correctly classified as Normal. | |
Model accuracy on Dataset 1.
| Train | Valid | Test | Train time | Test time | |
|---|---|---|---|---|---|
| InceptionV3 | 99.71 ± 0.11% | 96.55 ± 0.54% | 96.50 ± 0.82% | 00:14:27 | 00:00:03 |
| ResNet50V2 | 99.65 ± 0.25% | 95.51 ± 0.99% | 95.10 ± 1.93% | 00:10:40 | 00:00:02 |
| Xception | 99.77 ± 0.14% | 96.32 ± 0.50% | 95.89 ± 0.35% | 00:33:03 | 00:00:02 |
| DenseNet121 | 98.45 ± 0.45% | 95.14 ± 0.70% | 94.67 ± 1.28% | 00:14:30 | 00:00:03 |
| MobileNetV2 | 99.56 ± 0.41% | 94.87 ± 1.01% | 96.36 ± 1.02% | 00:08:47 | 00:00:01 |
| EfficientNetB0 | 95.09 ± 1.11% | 94.40 ± 0.60% | 94.80 ± 1.68% | 00:16:14 | 00:00:02 |
| EfficientNetV2 | 99.75 ± 0.14% | 97.13 ± 0.87% | 94.81 ± 0.58% | 00:21:32 | 00:00:03 |
| FT-InceptionV3 | 99.75 ± 0.11% | 98.10 ± 0.66% | 97.72 ± 0.38% | 00:12:23 | 00:00:03 |
| FT-ResNet50V2 | 99.88 ± 0.07% | 97.62 ± 0.64% | 97.13 ± 0.86% | 00:09:20 | 00:00:02 |
| FT-Xception | 99.86 ± 0.06% | 97.94 ± 0.47% | 97.31 ± 1.56% | 00:25:25 | 00:00:02 |
| FT-DenseNet121 | 99.90 ± 0.07% | 98.05 ± 0.60% | 95.41 ± 1.25% | 00:12:35 | 00:00:03 |
| FT-MobileNetV2 | 99.68 ± 0.32% | 97.16 ± 0.60% | 97.71 ± 1.31% | 00:05:21 | 00:00:01 |
| FT-EfficientNet-B0 | 99.13 ± 0.43% | 97.32 ± 0.59% | 96.77 ± 1.18% | 00:12:22 | 00:00:02 |
| FT-EfficientNetV2 | 99.78 ± 0.13% | 96.65 ± 1.14% | 96.73 ± 1.19% | 00:09:29 | 00:00:02 |
| LightEfficientNetV2 | 99.80 ± 0.16% | 95.17 ± 0.74% | 98.33 ± 0.99% | 00:04:30 | 00:00:01 |
Model performance on Dataset 1.
| Precision | Recall | F1-score | |
|---|---|---|---|
| InceptionV3 | 96.73 ± 0.83% | 96.77 ± 0.90% | 96.66 ± 0.83% |
| ResNet50V2 | 95.19 ± 1.87% | 95.10 ± 1.95% | 95.13 ± 1.82% |
| Xception | 96.54 ± 0.73% | 96.04 ± 0.45% | 96.50 ± 0.76% |
| DenseNet121 | 94.74 ± 1.45% | 94.63 ± 1.33% | 95.01 ± 1.29% |
| MobileNetV2 | 96.30 ± 0.99% | 96.33 ± 1.01% | 96.04 ± 0.80% |
| EfficientNetB0 | 94.54 ± 1.43% | 94.49 ± 1.51% | 94.46 ± 1.49% |
| EfficientNetV2 | 94.77 ± 0.57% | 94.83 ± 0.59% | 94.63 ± 0.39% |
| FT-InceptionV3 | 97.66 ± 0.46% | 97.69 ± 0.40% | 97.78 ± 0.58% |
| FT-ResNet50V2 | 96.77 ± 0.86% | 96.89 ± 1.06% | 96.87 ± 0.81% |
| FT-Xception | 97.46 ± 1.62% | 97.35 ± 1.59% | 97.24 ± 2.18% |
| FT-DenseNet121 | 95.14 ± 1.11% | 95.26 ± 1.31% | 95.19 ± 0.77% |
| FT-MobileNetV2 | 97.56 ± 1.46% | 97.65 ± 1.47% | 97.44 ± 1.48% |
| FT-EfficientNet-B0 | 96.71 ± 1.00% | 96.60 ± 1.10% | 96.63 ± 1.13% |
| FT-EfficientNetV2 | 96.81 ± 1.12% | 96.84 ± 1.12% | 96.70 ± 1.18% |
| LightEfficientNetV2 | 98.22 ± 0.88% | 98.30 ± 0.96% | 98.22 ± 0.90% |
Performance comparison on dataset 1.
| Models | Dataset 1 | |||
|---|---|---|---|---|
| Accuracy (%) | Precision (%) | Recall (%) | F1-score (%) | |
| FT-InceptionV3 | 97.72(±0.38)b | 97.66(±0.46)d | 97.69(±0.40)e | 97.78(±0.58)g |
| FT-ResNet50V2 | 97.13(±0.86) | 96.77(±0.86) | 96.89(±1.06) | 96.87(±0.81) |
| FT-Xception | 97.31(±1.56) | 97.46(±1.62) | 97.35(±1.59) | 97.24(±2.18) |
| FT-DenseNet121 | 95.41(±1.25)b,c | 95.14(±1.11)d | 95.26(±1.31)e,f | 95.19(±0.77)g |
| FT-MobileNetV2 | 97.71(±1.31)c | 97.56(±1.46) | 97.65(±1.47)f | 97.44(±1.48) |
| FT-EfficientNet-B0 | 96.77(±1.18) | 96.71(±1.00) | 96.60(±1.10) | 96.63(±1.13) |
| FT-EfficientNetV2 | 96.73(±1.19) | 96.81(±1.12) | 96.84(±1.12) | 96.70(±1.18) |
| LightEfficientNetV2 | 98.33(±0.99)a | 98.22(±0.88)a | 98.30(±0.96)a | 98.22(±0.90)a |
| P-value | 0.00** | 0.00** | 0.00** | 0.00** |
**p < 0.01.
a Means are statistically significant from the others.
b-g Means that share a letter are significantly different.
Model accuracy on Dataset 2.
| Train | Valid | Test | Train time | Test time | |
|---|---|---|---|---|---|
| InceptionV3 | 99.84 ± 0.10% | 92.53 ± 1.35% | 93.02 ± 1.15% | 00:14:31 | 00:00:03 |
| ResNet50V2 | 99.88 ± 0.10% | 93.65 ± 0.85% | 93.43 ± 1.91% | 00:10:33 | 00:00:02 |
| Xception | 98.20 ± 1.59% | 93.41 ± 1.25% | 93.46 ± 0.80% | 00:33:06 | 00:00:02 |
| DenseNet121 | 97.61 ± 0.88% | 93.68 ± 1.30% | 93.12 ± 2.21% | 00:14:33 | 00:00:03 |
| MobileNetV2 | 99.77 ± 0.13% | 94.36 ± 0.59% | 94.46 ± 1.44% | 00:08:49 | 00:00:01 |
| EfficientNetB0 | 94.51 ± 1.59% | 92.50 ± 1.16% | 93.40 ± 2.17% | 00:17:32 | 00:00:02 |
| EfficientNetV2 | 99.81 ± 0.07% | 93.46 ± 0.37% | 93.42 ± 0.31% | 00:21:37 | 00:00:03 |
| FT-InceptionV3 | 99.46 ± 0.35% | 94.70 ± 0.45% | 94.51 ± 0.76% | 00:12:20 | 00:00:03 |
| FT-ResNet50V2 | 99.84 ± 0.06% | 95.64 ± 0.54% | 96.01 ± 1.83% | 00:09:21 | 00:00:02 |
| FT-Xception | 99.88 ± 0.08% | 96.01 ± 0.67% | 96.78 ± 1.19% | 00:25:45 | 00:00:02 |
| FT-DenseNet121 | 99.88 ± 0.05% | 96.35 ± 1.03% | 96.58 ± 1.07% | 00:12:40 | 00:00:03 |
| FT-MobileNetV2 | 99.87 ± 0.09% | 95.58 ± 1.10% | 95.82 ± 1.42% | 00:05:24 | 00:00:01 |
| FT-EfficientNet-B0 | 99.44 ± 0.16% | 95.67 ± 0.39% | 95.73 ± 1.05% | 00:12:08 | 00:00:02 |
| FT-EfficientNetV2 | 99.82 ± 0.09% | 95.11 ± 0.83% | 96.67 ± 1.35% | 00:09:27 | 00:00:02 |
| LightEfficientNetV2 | 99.89 ± 0.08% | 96.65 ± 0.56% | 00:04:30 | 00:00:01 |
Model performance on Dataset 2.
| Precision | Recall | F1-score | |
|---|---|---|---|
| InceptionV3 | 93.20 ± 1.02% | 92.86 ± 1.02% | 92.83 ± 0.98% |
| ResNet50V2 | 93.61 ± 2.07% | 93.56 ± 2.00% | 93.46 ± 2.07% |
| Xception | 94.09 ± 0.82% | 94.06 ± 0.90% | 93.89 ± 0.77% |
| DenseNet121 | 93.11 ± 2.25% | 93.07 ± 2.23% | 93.06 ± 2.31% |
| MobileNetV2 | 94.51 ± 1.40% | 94.45 ± 1.40% | 94.41 ± 1.36% |
| EfficientNetB0 | 93.31 ± 2.20% | 93.30 ± 2.10% | 93.15 ± 2.20% |
| EfficientNetV2 | 93.46 ± 0.22% | 93.48 ± 0.24% | 93.34 ± 0.34% |
| FT-InceptionV3 | 94.77 ± 1.00% | 94.60 ± 0.85% | 94.57 ± 1.17% |
| FT-ResNet50V2 | 95.67 ± 1.90% | 95.67 ± 1.86% | 95.66 ± 2.07% |
| FT-Xception | 96.43 ± 1.15% | 96.52 ± 1.28% | 96.48 ± 1.35% |
| FT-DenseNet121 | 96.59 ± 1.02% | 96.53 ± 1.09% | 96.48 ± 1.18% |
| FT-MobileNetV2 | 95.74 ± 1.22% | 95.83 ± 1.47% | 95.06 ± 0.67% |
| FT-EfficientNet-B0 | 95.48 ± 0.23% | 95.74 ± 1.01% | 95.55 ± 0.74% |
| FT-EfficientNetV2 | 96.11 ± 1.30% | 96.09 ± 1.38% | 95.98 ± 1.35% |
| LightEfficientNetV2 | 97.41 ± 1.03% | 97.40 ± 0.96% | 97.49 ± 0.92% |
Performance comparison on dataset 2.
| Models | Dataset 2 | |||
|---|---|---|---|---|
| Accuracy (%) | Precision (%) | Recall (%) | F1-score (%) | |
| FT-InceptionV3 | 94.51(±0.76)b,c,d | 94.77(±1.00)e | 94.60(±0.85)f,g | 94.57(±1.17)h,i |
| FT-ResNet50V2 | 96.01(±1.83) | 95.67(±1.90) | 95.67(±1.86) | 95.66(±2.07) |
| FT-Xception | 96.78(±1.19)b | 96.43(±1.15) | 96.52(±1.28)g | 96.48(±1.35)i |
| FT-DenseNet121 | 96.58(±1.07)c | 96.59(±1.02)e | 96.53(±1.09)f | 96.48(±1.18)h |
| FT-MobileNetV2 | 95.82(±1.42) | 95.74(±1.22) | 95.83(±1.47) | 95.06(±0.67) |
| FT-EfficientNet-B0 | 95.73(±1.05) | 95.48(±0.23) | 95.74(±1.01) | 95.55(±0.74) |
| FT-EfficientNetV2 | 96.67(±1.35)d | 96.11(±1.30) | 96.09(±1.38) | 95.98(±1.35) |
| LightEfficientNetV2 | 97.48(±0.85)a | 97.41(±1.03)a | 97.40(±0.96)a | 97.49(±0.92)a |
| P-value | 0.00** | 0.00** | 0.00** | 0.00** |
**p < 0.01.
aMeans are statistically significant from the others.
b-g Means that share a letter are significantly different.
Comparison of the proposed model with SOTA.
| Study(s) | Dataset | Architecture | Class | Accuracy | Parameter |
|---|---|---|---|---|---|
| Panwar et al. (2020) [ | Total: 5863 X-ray images. | nCOVnet | 2 | 88.00% | 131,842 |
| Das et al. (2021) [ | 219 COVID-19, | TLCoV + VGG16 | 3 | 97.67% | 12,410,021 |
| Tahir et al. (2021) [ | 11596 COVID-19, | FPN + InceptionV4 | 3 | 99.23% | 4,357,000 |
| Sheykhivand et al. (2021) [ | 2842 COVID-19, | Proposed DNN | 4 | 99.50% | 23,070,232 |
| Saha et al. (2021) [ | 2300 COVID-19, | EMCNet | 2 | 98.91% | 3,955,009 |
| Ibrahim et al. (2021) [ | Total: 33676 (X-ray + CT images) | VGG19+ CNN | 4 | 98.05% | 22,337,604 |
| Hussain et al. (2021) [ | 500 COVID-19, | CoroDet | 3 | 94.20% | 2,873,609 |
| Panwar et al. (2020) [ | Total: 5856 X-ray images. | Proposed CNN | 2 | 95.61% | 5,244,098 |
| Total: 3008 CT images. | |||||
| Moghaddam et al. (2021) [ | 4001 COVID-19, | WCNN4 | 3 | 99.03% | 4,610,531 |
| Ahamed et al. (2021) [ | 1143 covid-19, | Modified & Tuned ResNet50V2 | 4 | 96.45% | 49,210,756 |
| 1000 covid-19, | 3 | 99.01% | |||
| Proposed | X-ray images | FT-EfficientNet-B0 | 3 | 96.77% | 912,123 |
| FT-DenseNet121 | 95.41% | 948,611 | |||
| FT-ResNet50V2 | 97.13% | 957,827 | |||
| FT-MobileNetV2 | 97.71% | 890,851 | |||
| FT-InceptionV3 | 97.81% | 913,299 | |||
| FT-Xception | 97.31% | 925,083 | |||
| FT-EfficientNetV2 | 96.73% | 863,895 | |||
| LightEfficientNetV2 | 98.33% | 798,539 | |||
| CT images | FT-EfficientNet-B0 | 3 | 95.73% | 912,123 | |
| FT-DenseNet121 | 96.58% | 948,611 | |||
| FT-ResNet50V2 | 96.01% | 957,827 | |||
| FT-MobileNetV2 | 95.82% | 890,851 | |||
| FT-InceptionV3 | 94.51% | 913,299 | |||
| FT-Xception | 96.78% | 925,083 | |||
| FT-EfficientNetV2 | 96.00% | 863,895 | |||
| LightEfficientNetV2 | 97.48% | 798,539 |
Comparisons on NIH Chest X-rays, SARS-CoV-2 and COVID-CT datasets.
| Dataset | Study(s) | Accuracy | Precision | Recall | F1-score | Parameter |
|---|---|---|---|---|---|---|
| NIH Chest X-rays | Wang et al. (2017) [ | 74.50% | – | – | – | – |
| Yao et al. (2018) [ | 76.10% | – | – | – | – | |
| Guendel et al. (2018) [ | 80.70% | – | – | – | – | |
| Li et al. (2018) [ | 75.50% | – | – | – | – | |
| Shen and Gao (2018) [ | 77.50% | – | – | – | – | |
| Tang et al. (2018) [ | 80.30% | – | – | – | – | |
| Liu et al. (2019) [ | 81.50% | – | – | – | – | |
| Bharati et al. (2020) [ | 73.00% | 69.00% | 63.00% | 68.00% | 15,488,051 | |
| Guan and Huang (2020) [ | 81.60% | – | – | – | – | |
| Proposed (LightEfficientNetV2) | ||||||
| SARS-CoV-2 | Özkaya et al. (2019) [ | 96.00% | – | – | – | – |
| Soares et al. (2020) [ | 97.30% | 99.10% | 95.50% | 97.30% | – | |
| Panwar et al. (2020) [ | 95.00% | 95.30% | 94.00% | 94.30% | 5,244,098 | |
| Jaiswal et al. (2021) [ | 96.20% | – | – | – | 20,242,984 | |
| Shaik and Cherukuri (2021) [ | 98.99% | 98.98% | 99.00% | 98.99% | – | |
| Proposed (LightEfficientNetV2) | ||||||
| COVID-CT | Mishra et al. (2020) [ | 88.30% | – | – | 86.70% | – |
| Saqib et al. (2020) [ | 80.30% | 78.20% | 85.70% | 81.80% | – | |
| He et al. (2020) [ | 86.00% | – | – | 85.00% | 14,149,480 | |
| Mobiny et al. (2020) [ | – | 84.00% | – | – | – | |
| Polsinelli et al. (2020) [ | 85.00% | 85.00% | 87.00% | 86.00% | 12,600,000 | |
| Yang et al. (2020) [ | – | – | 89.10% | – | 25,600,000 | |
| Cruz (2021) [ | 86.00% | – | 89.00% | 85.00% | – | |
| Shaik and Cherukuri (2021) [ | – | |||||
| Proposed (LightEfficientNetV2) | 88.67% | 87.28% | 87.43% | 87.55% |