| Literature DB >> 36061418 |
Ekram Chamseddine1, Nesrine Mansouri2, Makram Soui3, Mourad Abed4.
Abstract
Healthcare systems worldwide have been struggling since the beginning of the COVID-19 pandemic. The early diagnosis of this unprecedented infection has become their ultimate objective. Detecting positive patients from chest X-ray images is a quick and efficient solution for overloaded hospitals. Many studies based on deep learning (DL) techniques have shown high performance in classifying COVID-19 chest X-ray images. However, most of these studies suffer from a class imbalance problem mainly due to the limited number of COVID-19 samples. Such a problem may significantly reduce the efficiency of DL classifiers. In this work, we aim to build an accurate model that assists clinicians in the early diagnosis of COVID-19 using balanced data. To this end, we trained six state-of-the-art convolutional neural networks (CNNs) via transfer learning (TL) on three different COVID-19 datasets. The models were developed to perform a multi-classification task that distinguishes between COVID-19, normal, and viral pneumonia cases. To address the class imbalance issue, we first investigated the Weighted Categorical Loss (WCL) and then the Synthetic Minority Oversampling Technique (SMOTE) on each dataset separately. After a comparative study of the obtained results, we selected the model that achieved high classification results in terms of accuracy, sensitivity, specificity, precision, F1 score, and AUC compared to other recent works. DenseNet201 and VGG-19 claimed the best scores. With an accuracy of 98.87%, an F1_Score of 98.21%, a sensitivity of 98.86%, a specificity of 99.43%, a precision of 100%, and an AUC of 99.15%, the WCL combined with CheXNet outperformed the other examined models.Entities:
Keywords: COVID-19 diagnosis; Chest X-ray images; Classification; Data Imbalance; Deep learning; Transfer learning
Year: 2022 PMID: 36061418 PMCID: PMC9422401 DOI: 10.1016/j.asoc.2022.109588
Source DB: PubMed Journal: Appl Soft Comput ISSN: 1568-4946 Impact factor: 8.263
Fig. 1New confirmed deaths worldwide were caused by COVID-19 in March 2021 [1].
Fig. 2Typical CNN architecture [64].
Fig. 3The architecture of transfer learning inspired by [75].
Fig. 4An illustration of SMOTE oversampling technique.
Fig. 5An overview of the proposed approach.
The size of the training sets for dataset 1 and dataset 2 before and after data augmentation.
| Datasets | Class | Before DA | After DA |
|---|---|---|---|
| Dataset 1 | COVID-19 | 155 | 465 |
| Viral pneumonia | 1085 | 3255 | |
| Normal | 1085 | 3255 | |
| Dataset 2 | COVID-19 | 200 | 600 |
| Viral pneumonia | 950 | 2850 | |
| Normal | 950 | 2850 | |
| Dataset 3 | COVID-19 | 800 | – |
| Viral pneumonia | 5000 | – | |
| Normal | 5000 | – | |
Fig. 6(a) is the original image and (b) is the image with Contrast Limited Adaptive Histogram Equalization (CLAHE).
Classification results: without balancing technique.
| Dataset | CNN | ACC (%) | SEN (%) | SPE (%) | PPV (%) | NPV (%) | F1 (%) | AUC (%) |
|---|---|---|---|---|---|---|---|---|
| Dataset 1 | ||||||||
| CheXNet | 96.15 | 96.14 | 98.07 | 96.15 | 94.18 | 96.14 | 97.01 | |
| Xception | 97.05 | 97.05 | 98.53 | 97.05 | 96.65 | 97.05 | 97.10 | |
| ResNet152 | 95.46 | 95.46 | 95.50 | 95.46 | 94.43 | 95.48 | 96.60 | |
| MobileNetV2 | 94.10 | 94.08 | 97.16 | 94.13 | 93.21 | 94.10 | 95.58 | |
| Dataset 2 | ||||||||
| CheXNet | 90.49 | 90.43 | 95.47 | 90.50 | 82.90 | 90.40 | 93.40 | |
| Xception | 88.94 | 88.91 | 93.96 | 88.90 | 90.42 | 91.44 | 92.34 | |
| ResNet152 | 90.19 | 90.20 | 95.59 | 90.18 | 89.92 | 90.20 | 92.52 | |
| MobileNetV2 | 89.94 | 89.90 | 94.49 | 90.28 | 87.58 | 90.10 | 92.42 | |
| Dataset 3 | ||||||||
| CheXNet | 89.38 | 89.34 | 94.88 | 86.38 | 84.66 | 88.72 | 88.00 | |
| Xception | 91.44 | 91.42 | 95.72 | 91.40 | 89.56 | 91.44 | 88.61 | |
| ResNet152 | 91.89 | 91.96 | 96.94 | 92.00 | 90.05 | 91.91 | 88.06 | |
| MobileNetV2 | 91.44 | 91.51 | 95.72 | 91.44 | 90.07 | 91.57 | 88.45 | |
Classification results of experiment 1: Using weighted class balancing technique.
| Dataset | CNN | ACC (%) | SEN (%) | SPE (%) | PPV (%)(%) | NPV (%) | F1 (%) | AUC (%) |
|---|---|---|---|---|---|---|---|---|
| Dataset 1 | DenseNet201 121 | 96.60 | 96.59 | 98.29 | 96.58 | 94.58 | 96.36 | 97.80 |
| 100.00 | 98.61 | |||||||
| Xception | 95.92 | 95.91 | 97.95 | 95.92 | 92.90 | 95.91 | 96.94 | |
| ResNet152 | 97.51 | 97.50 | 98.75 | 97.51 | 95.22 | 97.50 | 98.13 | |
| 98.41 | 96.43 | |||||||
| MobileNetV2 | 95.69 | 95.69 | 97.84 | 95.69 | 93.90 | 95.69 | 96.76 | |
| Dataset 2 | 93.58 | 90.18 | ||||||
| CheXNet | 91.82 | 91.82 | 95.91 | 97.27 | 95.70 | 73.00 | 94.09 | |
| Xception | 89.94 | 89.91 | 94.96 | 90.05 | 88.15 | 89.55 | 92.34 | |
| ResNet152 | 87.55 | 87.56 | 93.77 | 89.29 | 90.29 | 90.86 | 93.81 | |
| 92.92 | 91.86 | |||||||
| MobileNetV2 | 91.57 | 91.57 | 95.78 | 91.57 | 90.17 | 91.68 | 93.86 | |
| Dataset 3 | 93.94 | 92.04 | ||||||
| CheXNet | 92.97 | 92.96 | 96.48 | 92.96 | 90.41 | 92.93 | 93.06 | |
| Xception | 93.00 | 92.91 | 96.45 | 93.00 | 90.12 | 92.87 | 93.62 | |
| ResNet152 | 90.43 | 90.43 | 95.21 | 90.42 | 89.37 | 90.37 | 89.00 | |
| 94.92 | 92.98 | |||||||
| MobileNetV2 | 95.69 | 91.51 | 95.51 | 91.62 | 88.53 | 91.57 | 90.69 | |
A summary of the proposed models and training hyperparameters with best results.
| Dataset | Models | Loss function | Activation function | Classifier | Optimizer (lr:learning rate) | Batch | Epochs |
|---|---|---|---|---|---|---|---|
| Dataset 1 | WCL | WCL | RELU | Softmax | Adam (lr | 32 | 50 |
| SMOTE | Categorical cross-entropy | RELU | Softmax | Adam (lr | 8 | 72 | |
| Dataset 2 | WCL | WCL | RELU | Softmax | Adam (lr | 8 | 20 |
| SMOTE | Categorical cross-entropy | RELU | Softmax | Adam (lr | 8 | 50 | |
| Dataset 3 | WCL | WCL | RELU | Softmax | Adam (lr | 8 | 20 |
| SMOTE | Categorical cross-entropy | RELU | Softmax | Adam (lr | 8 | 50 | |
Fig. 7Comparison of AUC metric provided by DenseNet201 for each dataset: (Left bar) without data balancing, (middle bar) with WCL and (right bar) with SMOTE.
Fig. 8Comparison of AUC metric provided by VGG19 for each dataset: (Left bar) without data balancing, (middle bar) with WCL, and (right bar) with SMOTE.
Classification results of experiment 2: Using SMOTE balancing technique.
| Dataset | CNN | ACC (%) | SEN (%) | SPE (%) | PPV (%) | NPV (%) | F1 (%) | AUC (%) |
|---|---|---|---|---|---|---|---|---|
| Dataset 1 | 98.64 | 98.64 | ||||||
| CheXNet | 97.51 | 97.50 | 98.75 | 97.51 | 97.51 | 97.50 | 98.20 | |
| Xception | 96.15 | 96.15 | 98.07 | 96.15 | 96.15 | 96.14 | 96.94 | |
| ResNet152 | 95.46 | 95.64 | 97.73 | 95.46 | 95.46 | 95.46 | 96.60 | |
| 98.19 | 98.19 | |||||||
| MobileNetV2 | 97.51 | 97.51 | 98.75 | 97.54 | 97.54 | 97.50 | 98.13 | |
| Dataset 2 | 92.20 | 92.20 | ||||||
| CheXNet | 91.19 | 91.19 | 95.59 | 91.18 | 91.18 | 91.12 | 93.42 | |
| Xception | 89.44 | 89.43 | 94.71 | 89.43 | 89.43 | 89.43 | 91.81 | |
| ResNet152 | 91.82 | 91.82 | 95.91 | 91.82 | 91.82 | 91.82 | 93.84 | |
| 92.08 | 92.08 | |||||||
| MobileNetV2 | 83.31 | 89.30 | 94.65 | 89.41 | 89.41 | 89.29 | 92.16 | |
| Dataset 3 | 92.77 | 92.77 | ||||||
| CheXNet | 92.02 | 92.02 | 92.02 | 96.01 | 96.01 | 92.10 | 89.54 | |
| Xception | 90.00 | 90.01 | 95.00 | 90.00 | 90.00 | 90.02 | 86.53 | |
| ResNet152 | 91.32 | 91.32 | 95.66 | 91.32 | 91.32 | 91.32 | 88.93 | |
| 93.66 | 93.66 | |||||||
| MobileNetV2 | 91.38 | 91.38 | 95.69 | 91.65 | 91.65 | 91.35 | 89.00 | |
Comparative analysis of the proposed model with recently proposed models.
| Dataset | Author | Model | ACC (%) | SEN (%) | SPE (%) | PPV (%) | F1 (%) | AUC (%) |
|---|---|---|---|---|---|---|---|---|
| Dataset 1 | Chowdhury et al. | DenseNet201 | 97.9 | 97.9 | 98.8 | 97.95 | – | – |
| Bassi et al. | DenseNet121 | 98.3 | – | – | 98.3 | 98.30 | – | |
| 98.21 | ||||||||
| 98.64 | 98.36 | 99.35 | 98.64 | 98.97 | ||||
| Dataset 2 | Haghanifar et al. | CheXNet | – | – | – | – | 85 | – |
| Ozturk et al. | DarkNet | 87.02 | – | – | – | – | – | |
| Dataset 3 | Wang et al. | COVID-Net | 92.4 | 88 | – | 91 | – | – |
| Oh et al. | ResNet-18 | 91.9 | – | – | 76.9 | – | – | |