| Literature DB >> 35261555 |
Abstract
In such a brief period, the recent coronavirus (COVID-19) already infected large populations worldwide. Diagnosing an infected individual requires a Real-Time Polymerase Chain Reaction (RT-PCR) test, which can become expensive and limited in most developing countries, making them rely on alternatives like Chest X-Rays (CXR) or Computerized Tomography (CT) scans. However, results from these imaging approaches radiated confusion for medical experts due to their similarities with other diseases like pneumonia. Other solutions based on Deep Convolutional Neural Network (DCNN) recently improved and automated the diagnosis of COVID-19 from CXRs and CT scans. However, upon examination, most proposed studies focused primarily on accuracy rather than deployment and reproduction, which may cause them to become difficult to reproduce and implement in locations with inadequate computing resources. Therefore, instead of focusing only on accuracy, this work investigated the effects of parameter reduction through a proposed truncation method and analyzed its effects. Various DCNNs had their architectures truncated, which retained only their initial core block, reducing their parameter sizes to <1 M. Once trained and validated, findings have shown that a DCNN with robust layer aggregations like the InceptionResNetV2 had less vulnerability to the adverse effects of the proposed truncation. The results also showed that from its full-length size of 55 M with 98.67% accuracy, the proposed truncation reduced its parameters to only 441 K and still attained an accuracy of 97.41%, outperforming other studies based on its size to performance ratio.Entities:
Keywords: Coronavirus pneumonia; Covid-19; Deep convolutional neural networks; Deep learning; Medical image diagnosis; Model truncation
Year: 2022 PMID: 35261555 PMCID: PMC8893243 DOI: 10.1007/s11042-022-12484-0
Source DB: PubMed Journal: Multimed Tools Appl ISSN: 1380-7501 Impact factor: 2.577
Summary of recent related studies
| Model | Dataset | Performance | Params |
|---|---|---|---|
| COVNet (ResNet50 backbone) [ | CT Scans (COVID-19, CAP, Non-Pneumonia) | Sensitivity: 90% Specificity: 96% | 25.6 M |
| Modified-Inception [ | CT Scans (COVID-19, non-COVID-19) | Accuracy: 89.5% | ≈23 M |
| VGG19 with Standard-CNN [ | CT Scans and X-Ray (Normal, Pneumonia, COVID-19, Lung Cancer) | Accuracy: 98.05% | 22.3 M |
| COVID-Net [ | CXRs (Normal, CAP, COVID-19 | Accuracy: 93.3% | 11.75 M |
| Location-attention oriented Model (based on ResNet18) [ | CT Scans (COVID-19, Influenza-A, Viral Pneumonia) | Accuracy: 86.7% | 11.7 M |
| Truncated InceptionNet (3 blocks) [ | CXRs (COVID-19, CAP, TB (China), TB (USA) | Accuracy: 99.92% | 2.1 M |
| COVIDNet-CT [ | CT Scans (COVID-19, non-COVID-19) | Accuracy: 99.1% | 1.4 M |
| Lightweight CNN (based on SqueezeNet) [ | CT Scans (COVID-19, non-COVID-19) | Accuracy: 85.03% | 1.26 M |
| Fused-DenseNet-Tiny [ | CXRs (Normal, COVID-19, CAP) | Accuracy: 97.99% | 1.2 M |
*Arranged according to parameter size
Fig. 1Samples of chest x-rays and computed tomography scans from the curated dataset
Dataset specification and distribution
| Computed Tomography Scans | |||
|---|---|---|---|
| Data | Train (80%) | Validation (20%) | Total |
| CT-Normal | 5515 | 1378 | 6893 |
| CT-COVID-19 | 6075 | 1518 | 7593 |
| CT-Pneumonia | 2095 | 523 | 2618 |
| CXR-Normal | 2616 | 654 | 3270 |
| CXR-COVID-19 | 1025 | 256 | 1281 |
| CXR-Pneumonia | 3726 | 931 | 4657 |
Specifications of the base and truncated models
| Model | BLL | TLL | BFS | TFS | Last layer name |
|---|---|---|---|---|---|
| InceptionV3 | 313 | 41 | 2048 | 256 | “mixed0” |
| Xception | 134 | 16 | 2048 | 128 | “add_1” |
| ResNet50V2 | 192 | 17 | 2048 | 256 | “conv2_block1_out” |
| InceptionResNetV2 | 782 | 41 | 1536 | 320 | “mixed_5b” |
| DenseNet121 | 429 | 14 | 1024 | 96 | “conv2_block1_concat” |
| EfficientNetB0 | 240 | 46 | 1280 | 24 | “block2b_add” |
Comparison of parameter size before and after truncation
| Model | BPS | TPS |
|---|---|---|
| InceptionV3 | 23,851,784 | 429,440 |
| Xception | 22,910,480 | 55,712 |
| ResNet50V2 | 25,613,800 | 84,480 |
| InceptionResNetV2 | 55,873,736 | 441,920 |
| DenseNet121 | 8,062,504 | 55,488 |
| EfficientNetB0 | 5,330,571 | 24,345 |
Fig. 2The proposed transfer learning and fine-tuning method
Fig. 3Proposed fine-tuning replacement layers for the truncated models
Hyper-parameter configuration
| Hyper-Parameter | Value |
|---|---|
| Batch Size | 16 |
| Optimizer | Adam |
| Learning Rate | 0.0001 |
| Epochs | 25 |
| Dropout Rate | 0.5 |
Learning rate reduction on plateau callback configuration
| Parameter | Value |
|---|---|
| Monitor | “Validation accuracy” |
| Factor | 0.5 |
| Patience | 2 |
| Minimum LR | 0.000001 |
Fig. 4Results of learning curves from the full-length models. a DenseNet121. b EfficientNetB0. c InceptionResNetV2. d InceptionV3. e ResNet50V2. f Xception
Fig. 5Accuracy and loss curves after the employment of the proposed truncation method. a DenseNet121-Tr. b EfficientNetB0-Tr. c InceptionResNetV2-Tr. d InceptionV3-Tr. e ResNet50V2-Tr. f Xception-Tr
Fig. 6Classification results of the truncated models based on a confusion matrix
Fig. 7Receiver operating characteristic of the truncated models
Fig. 8Precision-recall curves of the truncated models
Fig. 9Results comparison between the full-length and truncated models based on the mean squared log error (lower the better)
Overall performance of the truncated models
| Model | ||||
|---|---|---|---|---|
| DenseNet121-Tr | 77.98% | 80.86% | 76.72% | 78.25% |
| EfficientNetB0-Tr | 86.03% | 88.44% | 86.45% | 87.34% |
| InceptionResNetV2-Tr | ||||
| InceptionV3-Tr | 97.36% | 97.46% | 97.40% | 97.43% |
| ResNet50V2-Tr | 86.12% | 87.77% | 86.78% | 87.25% |
| Xception-Tr | 78.82% | 81.11% | 80.43% | 80.25% |
Fig. 10Gradient-weighted class activation maps of the full-length baseline models: DenseNet121 (a), EfficientNetB0 (b), InceptionResNetV2 (c), InceptionV3 (d), ResNet50V2 (e), Xception (f)
Fig. 11Gradient weighted class activation maps of the truncated models: DenseNet121-Tr (a), EfficientNetB0-Tr (b), InceptionResNetV2-Tr (c), InceptionV3-Tr (d), ResNet50V2-Tr (e), Xception-Tr (f)
Fig. 12Overall comparison of cost-efficiency to performance ratio
Comparison of storage consumption to performance ratio between full-length and truncated models
| Model | Full-length | Truncated | ||
|---|---|---|---|---|
| Size in | Size in | |||
| DenseNet121 | 83,446 | 714 | 77.98% | |
| EfficientNetB0 | 98.73% | 86.03% | ||
| InceptionResNetV2 | 638,982 | 98.67% | 5338 | |
| InceptionV3 | 256,430 | 98.40% | 5188 | 97.36% |
| ResNet50V2 | 276,649 | 98.48% | 1081 | 86.12% |
| Xception | 244,766 | 98.57% | 739 | 78.82% |
Hyper-parameter settings across various combinations
| Description | Optimizer | LR | Dropout rate |
|---|---|---|---|
| Setting-1 (a) | SGD | 0.01 | 0.1 |
| Setting-1 (b) | 0.3 | ||
| Setting-1 (c) | 0.5 | ||
| Setting-1 (d) | No dropout (Untuned) | ||
| Setting-2 (a) | RMSprop | 0.001 | 0.1 |
| Setting-2 (b) | 0.3 | ||
| Setting-2 (c) | 0.5 | ||
| Setting-2 (d) | No dropout (Untuned) | ||
| Setting-3 (a) | Adam | 0.0001 | 0.1 |
| Setting-3 (b) | 0.3 | ||
| Setting-3 (c) | 0.5 | ||
| Setting-3 (d) | No dropout (Untuned) |
Setting-1 results
| Model | Overall accuracy (%) | |||
|---|---|---|---|---|
| DenseNet121-Tr | 65.8 | 64.33 | 55.78 | 70.78 |
| EfficientNetB0-Tr | 89.03 | 78.44 | 86.75 | 87.19 |
| InceptionV3-Tr | 92.36 | 91.52 | 93.94 | 89.39 |
| ResNet50V2-Tr | 77.83 | 65.27 | 51.81 | 39.81 |
| Xception-Tr | 72.22 | 77.03 | 59.41 | 79.87 |
| InceptionResNetV2-Tr | 90.78 | 93.16 | 91.05 | |
Setting-2 results
| Model | Overall accuracy (%) | |||
|---|---|---|---|---|
| DenseNet121-Tr | 77.15 | 80.99 | 77.62 | 82.53 |
| EfficientNetB0-Tr | 85.34 | 93.44 | 82.68 | 89.01 |
| InceptionV3-Tr | 86.84 | 65.36 | 91.71 | 92.36 |
| ResNet50V2-Tr | 70.87 | 85.89 | 46.44 | 74.92 |
| Xception-Tr | 81.37 | 74.37 | 77.41 | 86.14 |
| InceptionResNetV2-Tr | 96.88 | 93.4 | 95.57 | |
Setting-3 results
| Model | Overall accuracy (%) | |||
|---|---|---|---|---|
| DenseNet121-Tr | 79.07 | 80.06 | 77.98 | 76.5 |
| EfficientNetB0-Tr | 88.40 | 88.31 | 86.03 | 90.74 |
| InceptionV3-Tr | 94.54 | 96.58 | 97.36 | 93.99 |
| ResNet50V2-Tr | 89.01 | 85.10 | 86.12 | 80.10 |
| Xception-Tr | 77.30 | 82.36 | 78.82 | 82.38 |
| InceptionResNetV2-Tr | 95.99 | 96.73 | 95.29 | |
Comparison of the top-scoring truncated model against other existing studies
| Model | Dataset | Performance | Params |
|---|---|---|---|
| CXRs and CT scans (Normal, COVID-19, Pneumonia) | Accuracy: 97.41% | ||
| COVNet (ResNet50 backbone) [ | CT Scans (COVID-19, CAP, Non-Pneumonia) | Sensitivity: 90% Specificity: 96% | 25.6 M |
| Modified-Inception [ | CT Scans (COVID-19, non-COVID-19) | Accuracy: 89.5% | ≈23 M |
| VGG19 with Standard-CNN [ | CT Scans and X-Ray (Normal, Pneumonia, COVID-19, Lung Cancer) | Accuracy: 98.05% | 22.3 M |
| COVID-Net [ | CXRs (Normal, CAP, COVID-19 | Accuracy: 93.3% | 11.75 M |
| Location-attention oriented Model (based on ResNet18) [ | CT Scans (COVID-19, Influenza-A, Viral Pneumonia) | Accuracy: 86.7% | 11.7 M |
| Truncated InceptionNet (3 blocks) [ | CXRs (COVID-19, CAP, TB (China), TB (USA) | Accuracy: 99.92% | 2.1 M |
| COVIDNet-CT [ | CT Scans (COVID-19, non-COVID-19) | Accuracy: 99.1% | 1.4 M |
| Lightweight CNN (based on SqueezeNet) [ | CT Scans (COVID-19, non-COVID-19) | Accuracy: 85.03% | 1.26 M |
| Fused-DenseNet-Tiny [ | CXRs (Normal, COVID-19, CAP) | Accuracy: 97.99% | 1.2 M |