| Literature DB >> 35290811 |
Nadiah A Baghdadi1, Amer Malki2, Sally F Abdelaliem3, Hossam Magdy Balaha4, Mahmoud Badawy5, Mostafa Elhosseini6.
Abstract
Researchers have developed more intelligent, highly responsive, and efficient detection methods owing to the COVID-19 demands for more widespread diagnosis. The work done deals with developing an AI-based framework that can help radiologists and other healthcare professionals diagnose COVID-19 cases at a high level of accuracy. However, in the absence of publicly available CT datasets, the development of such AI tools can prove challenging. Therefore, an algorithm for performing automatic and accurate COVID-19 classification using Convolutional Neural Network (CNN), pre-trained model, and Sparrow search algorithm (SSA) on CT lung images was proposed. The pre-trained CNN models used are SeresNext50, SeresNext101, SeNet154, MobileNet, MobileNetV2, MobileNetV3Small, and MobileNetV3Large. In addition, the SSA will be used to optimize the different CNN and transfer learning(TL) hyperparameters to find the best configuration for the pre-trained model used and enhance its performance. Two datasets are used in the experiments. There are two classes in the first dataset, while three in the second. The authors combined two publicly available COVID-19 datasets as the first dataset, namely the COVID-19 Lung CT Scans and COVID-19 CT Scan Dataset. In total, 14,486 images were included in this study. The authors analyzed the Large COVID-19 CT scan slice dataset in the second dataset, which utilized 17,104 images. Compared to other pre-trained models on both classes datasets, MobileNetV3Large pre-trained is the best model. As far as the three-classes dataset is concerned, a model trained on SeNet154 is the best available. Results show that, when compared to other CNN models like LeNet-5 CNN, COVID faster R-CNN, Light CNN, Fuzzy + CNN, Dynamic CNN, CNN and Optimized CNN, the proposed Framework achieves the best accuracy of 99.74% (two classes) and 98% (three classes).Entities:
Keywords: COVID-19; Convolutional neural network (CNN); Deep learning (DL); Metaheuristic optimization; Sparrow search algorithm
Mesh:
Year: 2022 PMID: 35290811 PMCID: PMC8906898 DOI: 10.1016/j.compbiomed.2022.105383
Source DB: PubMed Journal: Comput Biol Med ISSN: 0010-4825 Impact factor: 4.589
Fig. 1Worldwide daily new confirmed COVID-19 cases and deaths per million people [4].
Fig. 2COVID-19 cases reported weekly by WHO Region, and global deaths, as of 2 January 2022 [3].
Fig. 3COVID-19 diagnosis techniques.
Fig. 4CCT scan of lungs of a patient (a) affected by COVID-19 and (b) not affected by COVID-19.
Fig. 5The suggested framework.
The configurations used for the different augmentation techniques to balance the datasets.
| Technique | Value |
|---|---|
| Rotation | 30° |
| Width Shift Ratio | 20% |
| Height Shift Ratio | 20% |
| Shear Ratio | 20% |
| Zoom Ratio | 20% |
| Brightness change | [0.8, 1.2] |
| Vertical Flip | Yes |
| Horizontal Flip | Yes |
The corresponding hyperparameter for each element in the solution.
| Element # | Value |
|---|---|
| 1 | Loss function |
| 2 | Batch size |
| 3 | Dropout ratio |
| 4 | TL learning ratio |
| 5 | Weights optimizer |
| 6 | Scaler technique |
| 7 | Apply augmentation or not |
| 8 | Rotation value (if augmentation is true) |
| 9 | Width shift value (if augmentation is true) |
| 10 | Height shift value (if augmentation is true) |
| 11 | Shear value (if augmentation is true) |
| 12 | Zoom value (if augmentation is true) |
| 13 | Horizontal flip flag (if augmentation is true) |
| 14 | Vertical flip flag (if augmentation is true) |
| 15 | Brightness change range (if augmentation is true) |
Common experiments Configurations.
| Configuration | Specifications |
|---|---|
| Apply Dataset Shuffling? | Yes (Random) |
| Input Image Size | (100 × 100 × 3) |
| Hyperparameters Metaheuristic Optimizer | Sparrow Search Algorithm (SpaSA) |
| Train Split Ratio | 85%–15% (i.e., 85% for training and validation; and 15% for testing) |
| SpaSA Size of Population | 10 |
| SpaSA Number of Iterations | 10 |
| Number of Epochs | 5 |
| Output Activation Function | SoftMax |
| Pre-trained Models | SeresNext50, SeresNext101, SeNet154, MobileNet, MobileNetV2, MobileNetV3Small, and MobileNetV3Large |
| Pre-trained Parameters Initializers | ImageNet |
| Losses Range | Categorical Crossentropy, Categorical Hinge, KLDivergence, Poisson, Squared Hinge, and Hinge |
| Parameters Optimizers Range | Adam, NAdam, AdaGrad, AdaDelta, AdaMax, RMSProp, SGD, Ftrl, SGD Nesterov, RMSProp Centered, and Adam AMSGrad |
| Dropout Range | [0 → 0.6] |
| Batch Size Range | 4 → 48 (step = 4) |
| Pre-trained Model Learn Ratio Range | 1 → 100 (step = 1) |
| Scaling Techniques | Normalize, Standard, Min Max, and Max Abs |
| Apply Data Augmentation (DA) | [ |
| DA Rotation Range | 0° → 45° (step = 1°) |
| DA Width Shift Range | [0 → 0.25] |
| DA Height Shift Range | [0 → 0.25] |
| DA Shear Range | [0 → 0.25] |
| DA Zoom Range | [0 → 0.25] |
| DA Horizontal Flip Range | [ |
| DA Vertical Flip Range | [ |
| DA Brightness Range | [0.5 → 2.0] |
| Scripting Language | Python |
| Python Major Packages | Tensorflow, Keras, NumPy, OpenCV, and Matplotlib |
| Working Environment | Google Colab with GPU (i.e., Intel(R) Xeon(R) CPU @ 2.00 GHz, Tesla T4 16 GB GPU, CUDA v.11.2, and 12 GB RAM) |
The used datasets specifications summarization.
| Dataset | No. of Classes | Classes | No. of Images (Before) | No. of Images (After) |
|---|---|---|---|---|
| 2 | “COVID” and “NonCOVID” | 14, 486 | 15, 186 | |
| 3 | “CAP”, “COVID”, and “NonCOVID” | 17, 104 | 22, 779 |
Fig. 6Samples from the used datasets.
Two-classes specific experiments Configurations.
| Configuration | Specifications |
|---|---|
| Dataset Sources | COVID-19 Lung CT Scans [ |
| Number of Classes | 2 |
| Classes | (‘COVID’ and ‘NonCOVID’) |
| Dataset Size before Data Balancing | “COVID”: 7,593 and “NonCOVID”: 6,893 |
| Dataset Size after Data Balancing | “COVID”: 7,593 and “NonCOVID”: 7,593 |
Confusion matrix results concerning the two-classes dataset.
| Model Name | TP | TN | FP | FN |
|---|---|---|---|---|
| SeresNext50 | 15,022 | 15,022 | 158 | 158 |
| SeresNext101 | 15,064 | 15,064 | 104 | 104 |
| SeNet154 | 14,966 | 14,966 | 214 | 214 |
| MobileNet | 15,141 | 15,141 | 39 | 39 |
| MobileNetV2 | 15,088 | 15,088 | 72 | 72 |
| MobileNetV3Small | 14,282 | 14,282 | 898 | 898 |
| MobileNetV3Large | 14,768 | 14,768 | 392 | 392 |
The best solutions after the learning and optimization process concerning the two-classes dataset.
| Model Name | Loss | Batch Size | Dropout | TL Learn Ratio | Optimizer | Scaler | Apply Augmentation | Rotation Range | Width Shift Range | Height Shift Range | Shear Range | Zoom Range | Horizontal Flip | Vertical Flip | Brightness Range |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| SeresNext50 | Categorical Crossentropy | 12 | 0.2 | 89 | SGD | Standardize | No | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A |
| SeresNext101 | KLDivergence | 24 | 0.07 | 29 | SGD Nesterov | MinMax | Yes | 16 | 0.25 | 0.23 | 0.05 | 0.05 | No | No | 1.2–1.87 |
| SeNet154 | Poisson | 44 | 0.22 | 63 | SGD | MinMax | Yes | 29 | 0.13 | 0.1 | 0.18 | 0 | No | No | 1.08–1.55 |
| MobileNet | KLDivergence | 44 | 0.37 | 60 | SGD | MaxAbs | No | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A |
| MobileNetV2 | KLDivergence | 40 | 0 | 62 | SGD Nesterov | MinMax | Yes | 36 | 0.04 | 0.09 | 0.11 | 0.09 | Yes | Yes | 1.32–1.79 |
| MobileNetV3Small | Squared Hinge | 12 | 0.23 | 100 | SGD | Standardize | Yes | 41 | 0.09 | 0.15 | 0.05 | 0.09 | No | No | 0.57–1.56 |
| MobileNetV3Large | KLDivergence | 40 | 0.1 | 45 | SGD | Standardize | Yes | 37 | 0.18 | 0.05 | 0.06 | 0.04 | Yes | Yes | 0.65–0.7 |
The two-classes dataset experiments with the maxmimized metrics.
| Model Name | Accuracy | F1 | Precision | Recall | Specificity | Sensitivity | AUC | IoU | Dice | Cosine Similarity |
|---|---|---|---|---|---|---|---|---|---|---|
| SeresNext50 | 98.96% | 98.96% | 98.96% | 98.96% | 98.96% | 98.96% | 99.89% | 98.40% | 98.72% | 99.09% |
| SeresNext101 | 97.41% | 97.41% | 97.41% | 97.41% | 97.41% | 97.41% | 99.68% | 96.61% | 97.25% | 97.84% |
| SeNet154 | 99.31% | 99.31% | 99.31% | 99.31% | 99.31% | 99.31% | 99.87% | 99.18% | 99.32% | 99.40% |
| MobileNet | 98.59% | 98.59% | 98.59% | 98.59% | 98.59% | 98.59% | 99.83% | 97.66% | 98.13% | 98.68% |
| MobileNetV2 | 94.08% | 94.08% | 94.08% | 94.08% | 94.08% | 94.08% | 97.81% | 95.01% | 95.52% | 94.78% |
| MobileNetV3Small | 99.53% | 99.53% | 99.53% | 99.53% | 99.53% | 99.53% | 99.96% | 98.89% | 99.15% | 99.54% |
| MobileNetV3Large | 99.74% | 99.74% | 99.74% | 99.74% | 99.74% | 99.74% | 99.97% | 99.69% | 99.74% | 99.78% |
The two-classes dataset experiments with the minimized metrics.
| Model Name | Categorical Crossentropy | KLDivergence | Categorical Hinge | Hinge | Squared Hinge | Poisson | Logcosh Error | Mean Absolute Error | Mean Squared Error | Mean Squared Logarithmic Error | Root Mean Squared Error |
|---|---|---|---|---|---|---|---|---|---|---|---|
| SeresNext50 | 0.033 | 0.033 | 0.038 | 0.519 | 0.528 | 0.517 | 0.004 | 0.019 | 0.009 | 0.004 | 0.092 |
| SeresNext101 | 0.069 | 0.069 | 0.083 | 0.541 | 0.561 | 0.534 | 0.009 | 0.041 | 0.020 | 0.010 | 0.140 |
| SeNet154 | 0.024 | 0.024 | 0.021 | 0.510 | 0.516 | 0.512 | 0.003 | 0.010 | 0.006 | 0.003 | 0.075 |
| MobileNet | 0.047 | 0.047 | 0.056 | 0.528 | 0.540 | 0.523 | 0.006 | 0.028 | 0.012 | 0.006 | 0.111 |
| MobileNetV2 | 0.229 | 0.229 | 0.134 | 0.567 | 0.616 | 0.614 | 0.022 | 0.067 | 0.049 | 0.024 | 0.221 |
| MobileNetV3Small | 0.019 | 0.019 | 0.026 | 0.513 | 0.517 | 0.510 | 0.002 | 0.013 | 0.004 | 0.002 | 0.067 |
| MobileNetV3Large | 0.008 | 0.008 | 0.008 | 0.504 | 0.506 | 0.504 | 0.001 | 0.004 | 0.002 | 0.001 | 0.045 |
Three-classes specific experiments Configurations.
| Configuration | Specifications |
|---|---|
| Dataset Source | Large COVID-19 CT scan slice dataset [ |
| Number of Classes | 3 |
| Classes | (‘COVID’, ‘NonCOVID’, and ‘CAP’) |
| Dataset Size before Data Balancing | “COVID”: 7,593, “NonCOVID”: 6,893, and “CAP”: 2,618 |
| Dataset Size after Data Balancing | “COVID”: 7,593, “NonCOVID”: 7,593, and “CAP”: 7,593 |
Confusion matrix results concerning the three-classes dataset.
| Model Name | TP | TN | FP | FN |
|---|---|---|---|---|
| SeresNext50 | 22,200 | 44,956 | 540 | 548 |
| SeresNext101 | 21,585 | 44,554 | 966 | 1,175 |
| SeNet154 | 21,312 | 44,136 | 1,384 | 1,448 |
| MobileNet | 22,299 | 45,074 | 446 | 461 |
| MobileNetV2 | 21,574 | 44,364 | 1,172 | 1,194 |
| MobileNetV3Small | 16,961 | 40,707 | 4,845 | 5,815 |
| MobileNetV3Large | 21,318 | 44,088 | 1,416 | 1,434 |
The best solutions after the learning and optimization process concerning the three-classes dataset.
| Model Name | Loss | Batch Size | Dropout | TL Learn Ratio | Optimizer | Scaler | Apply Augmentation | Rotation Range | Width Shift Range | Height Shift Range | Shear Range | Zoom Range | Horizontal Flip | Vertical Flip | Brightness Range |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| SeresNext50 | Poisson | 44 | 0.2 | 26 | AdaMax | MinMax | No | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A |
| SeresNext101 | Poisson | 20 | 0.45 | 52 | SGD Nesterov | MinMax | Yes | 23 | 0.15 | 0.02 | 0 | 0.01 | Yes | Yes | 0.57–1.25 |
| SeNet154 | Squared Hinge | 40 | 0 | 27 | AdaGrad | MinMax | Yes | 11 | 0.03 | 0.22 | 0.07 | 0.25 | Yes | No | 1.4–1.52 |
| MobileNet | Categorical Crossentropy | 20 | 0.08 | 75 | AdaMax | MaxAbs | Yes | 11 | 0.06 | 0.05 | 0.13 | 0.14 | Yes | No | 1.45–1.59 |
| MobileNetV2 | Squared Hinge | 16 | 0.53 | 63 | SGD Nesterov | MinMax | No | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A |
| MobileNetV3Small | Squared Hinge | 12 | 0.2 | 91 | AdaGrad | Normalize | No | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A |
| MobileNetV3Large | Categorical Crossentropy | 36 | 0.3 | 31 | AdaMax | Standardize | No | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A |
The three-classes dataset experiments with the maxmimized metrics.
| Model Name | Accuracy | F1 | Precision | Recall | Specificity | Sensitivity | AUC | IoU | Dice | Cosine Similarity |
|---|---|---|---|---|---|---|---|---|---|---|
| SeresNext50 | 95.25% | 95.21% | 95.65% | 94.84% | 97.88% | 94.84% | 99.54% | 94.55% | 95.44% | 96.12% |
| SeresNext101 | 97.61% | 97.61% | 97.63% | 97.59% | 98.81% | 97.59% | 99.83% | 97.31% | 97.75% | 98.02% |
| SeNet154 | 98.00% | 98.00% | 98.04% | 97.97% | 99.02% | 97.97% | 99.92% | 96.94% | 97.57% | 98.36% |
| MobileNet | 94.80% | 94.80% | 94.85% | 94.76% | 97.43% | 94.76% | 98.07% | 95.43% | 95.88% | 95.19% |
| MobileNetV2 | 76.15% | 75.92% | 77.90% | 74.47% | 89.36% | 74.47% | 88.59% | 76.67% | 79.68% | 79.66% |
| MobileNetV3Small | 93.70% | 93.76% | 93.89% | 93.64% | 96.96% | 93.64% | 97.43% | 94.31% | 94.92% | 94.29% |
| MobileNetV3Large | 93.73% | 93.73% | 93.77% | 93.70% | 96.89% | 93.70% | 98.20% | 94.71% | 95.25% | 94.47% |
The three-classes dataset experiments with the minimized metrics.
| Model Name | Categorical Crossentropy | KLDivergence | Categorical Hinge | Hinge | Squared Hinge | Poisson | Logcosh Error | Mean Absolute Error | Mean Squared Error | Mean Squared Logarithmic Error | Root Mean Squared Error |
|---|---|---|---|---|---|---|---|---|---|---|---|
| SeresNext50 | 0.123 | 0.123 | 0.129 | 0.712 | 0.735 | 0.374 | 0.011 | 0.046 | 0.023 | 0.011 | 0.151 |
| SeresNext101 | 0.065 | 0.065 | 0.067 | 0.689 | 0.701 | 0.355 | 0.006 | 0.022 | 0.012 | 0.006 | 0.110 |
| SeNet154 | 0.054 | 0.054 | 0.072 | 0.691 | 0.701 | 0.351 | 0.005 | 0.024 | 0.010 | 0.005 | 0.100 |
| MobileNet | 0.279 | 0.278 | 0.122 | 0.708 | 0.738 | 0.426 | 0.014 | 0.041 | 0.031 | 0.015 | 0.175 |
| MobileNetV2 | 0.793 | 0.793 | 0.567 | 0.870 | 0.989 | 0.598 | 0.054 | 0.203 | 0.119 | 0.059 | 0.345 |
| MobileNetV3Small | 0.398 | 0.386 | 0.148 | 0.717 | 0.753 | 0.461 | 0.016 | 0.051 | 0.036 | 0.017 | 0.189 |
| MobileNetV3Large | 0.294 | 0.285 | 0.141 | 0.714 | 0.749 | 0.428 | 0.015 | 0.048 | 0.035 | 0.017 | 0.186 |
Fig. 7Hyperparameters selection and best combinations graphical summarization.
Fig. 8Summarization of the learning and optimization experiments related to the two-classes dataset.
Fig. 9Summarization of the learning and optimization experiments related to the three-classes dataset.
A comparison between the data augmentation and train-to-test splitting approach and cross-validation approach.
| Approach | Accuracy | AUC | Cosine Similarity | TP | TN | FP | FN |
|---|---|---|---|---|---|---|---|
| Data augmentation and train-to-test splitting approach | 99.74% | 99.97% | 99.78% | 14,768 | 14,768 | 392 | 392 |
| Cross-validation approach | 84.99% | 92.60% | 87.72% | 2,581 | 2,581 | 455 | 455 |
A comparison between the optimized and non-optimized approaches.
| Approach | Accuracy | AUC | Cosine Similarity | TP | TN | FP | FN |
|---|---|---|---|---|---|---|---|
| Optimized Approach | 99.74% | 99.97% | 99.78% | 14,768 | 14,768 | 392 | 392 |
| Non-optimized Approach | 83.33% | 91.70% | 86.60% | 3,164 | 3,164 | 633 | 633 |
A comparison between the existence and non-existence of transfer learning.
| Approach | Accuracy | AUC | Cosine Similarity | TP | TN | FP | FN |
|---|---|---|---|---|---|---|---|
| With Transfer Learning | 99.74% | 99.97% | 99.78% | 14,768 | 14,768 | 392 | 392 |
| Without Transfer Learning | 49.67% | 70.70% | 49.70% | 1,886 | 1,886 | 1,911 | 1,911 |
Comparison between the suggested approach and related studies.
| Study | Year | Dataset | Approach | Best Accuracy |
|---|---|---|---|---|
| Islam et al. [ | 2020 | CCT | LeNet-5 CNN | 86.06% |
| Shibly et al. [ | 2020 | CXR | COVID faster R–CNN | 97.36% |
| Polsinelli et al. [ | 2020 | CCT | Light CNN | 85.03% |
| Tripti Goel et al. [ | 2020 | CCT | CNN + GAN | 99.22% |
| Huang et al. [ | 2020 | CCT | MCSL | 98.03% |
| Abraham and Nair [ | 2020 | CCT | CNN + KSVM | 91.60% |
| Kundu et al. [ | 2021 | CCT | Fuzzy + CNN | 98.93% and 98.80% |
| Jia et al. [ | 2021 | CXR and CCT | Dynamic CNN | 99.6% (CXR) and 99.3% (CCT) |
| Maghdid et al. [ | 2021 | CXR and CCT | CNN | 98% |
| Pathan et al. [ | 2021 | CCT | Optimized CNN | 98% |
| R. Murugan and Tripti Goel [ | 2021 | CXR | E-DiCoNet | 94.07% |
| Goura and Jain [ | 2022 | CCT + CXR | DLS-CNN | 98.78% |
| Gayathri et al. [ | 2022 | CXR | FFNN | 95.78% |
| Tripti Goel et al. [ | 2022 | CXR | MOGOA | 98.27% |
| Guoqing et al. [ | 2022 | CCT + CXR | COVID-MTL | 98.78% |
| Shaik and Cherukuri [ | 2022 | CCT | DNN | 93.33% |
| Current Study | 2022 | CT | Hybrid (SpaSA and CNN) | 99.74% (two-classes) and 98% (three-classes) |