| Literature DB >> 35941907 |
Abstract
Pneumonia is among the common symptoms of the virus that causes COVID-19, which has turned into a worldwide pandemic. It is possible to diagnose pneumonia by examining chest radiographs. Chest x-ray (CXR) is a fast, low-cost, and practical method widely used in this field. The fact that different pathogens other than COVID-19 also cause pneumonia and the radiographic images of all are similar make it difficult to detect the source of the disease. In this study, automatic detection of COVID-19 cases over CXR images was tried to be performed using convolutional neural network (CNN), a deep learning technique. Classifications were carried out using six different architectures on the dataset consisting of 15,153 images of three different types: healthy, COVID-19, and other viral-induced pneumonia. In the classifications performed with five different state-of-art models, ResNet18, GoogLeNet, AlexNet, VGG16, and DenseNet161, and a minimal CNN architecture specific to this study, the most successful result was obtained with the ResNet18 architecture as 99.25% accuracy. Although the minimal CNN model developed for this study has a simpler structure, it was observed that it has a success to compete with more complex models. The performances of the models used in this study were compared with similar studies in the literature and it was revealed that they generally achieved higher success. The model with the highest success was transformed into a test application, tested by 10 volunteer clinicians, and it was concluded that it provides 99.06% accuracy in practical use. This result reveals that the conducted study can play the role of a successful decision support system for experts.Entities:
Keywords: COVID‐19 diagnosis; chest x‐ray analysis; convolutional neural network; pneumonia detection
Year: 2022 PMID: 35941907 PMCID: PMC9348396 DOI: 10.1111/coin.12526
Source DB: PubMed Journal: Comput Intell ISSN: 0824-7935 Impact factor: 2.142
FIGURE 1COVID‐19 cases of the top 10 countries
FIGURE 2Class distributions of the dataset
FIGURE 3Sample images from the dataset
FIGURE 4The general structure of a CNN model
Advantages and disadvantages of the state‐of‐art models used in this study
| Model | Advantages | Disadvantages |
|---|---|---|
| ResNet18 |
Increases the depth of the network instead of widening it, so fewer additional parameters are required. The training process is faster. Reduces the effects of the vanishing gradient problem. |
Increased complexity. The model requires substantial Batch Normalization. Jump links need to be added between different layers. |
| GoogLeNet |
It is trained faster than VGG. The size of the trained model is slightly smaller. |
Although it has fewer parameters, the model structure is deeper and more complex. |
| AlexNet |
It has the ability to quickly subsample intermediate representations. Extracts features better than the LeNet model. Performs well on colored images. |
It has slightly less depth and therefore takes more effort to extract image features. Longer training is required to achieve higher accuracy. |
| VGG16 |
The model has enhanced depth that increases its success. More layers with smaller kernels increase non‐linearity. |
Network training takes longer. The weights in the network architecture are large. It is more prone to the vanishing gradient problem than ResNet. |
| DenseNet |
Alleviates the vanishing gradient problem. Strengthens feature propagation. Allows the reuse of features. |
Dense connections require more memory usage. Requires more training time. |
Encoded labels of each class
| Image label | Normal (healthy) | COVID‐19 | Other viral pneumonia |
|---|---|---|---|
| Encoded label | 0 | 1 | 2 |
FIGURE 5The general structure of the developed CNN model
Summary of the model
| Layer type | Layer features | Output shape | Parameters | |
|---|---|---|---|---|
| CNN model | Convolution 2D | Filters: 32, Kernel: 3 × 3, Activation: ReLU | (299, 299, 32) | 896 |
| Convolution 2D | Filters: 64, Kernel: 3 × 3, Activation: ReLU | (299, 299, 64) | 18,496 | |
| Max pooling 2D | Pool size: 2 × 2 | (149, 149, 64) | 0 | |
| Dropout | 25% | (149, 149, 64) | 0 | |
| Convolution 2D | Filters: 64, Kernel: 3 × 3, Activation: ReLU | (149, 149, 64) | 36,928 | |
| Max pooling 2D | Pool size: 2 × 2 | (74, 74, 64) | 0 | |
| Dropout | 25% | (74, 74, 64) | 0 | |
| Convolution 2D | Filters: 128, Kernel: 3 × 3, Activation: ReLU | (74, 74, 128) | 73,856 | |
| Max pooling 2D | Pool size: 2 × 2 | (37, 37, 128) | 0 | |
| Dropout | 25% | (37, 37, 128) | 0 | |
| Flatten | (175,232) | 0 | ||
| Classifier NN | Dense | Neurons: 256, Activation: ReLU | (256) | 40,141,056 |
| Dense | Neurons: 128, Activation: ReLU | (128) | 32,896 | |
| Dense | Neurons: 64, Activation: ReLU | (64) | 8256 | |
| Dropout | 50% | (None, 64) | 0 | |
| Dense | Neurons: 3, Activation: Softmax | (None, 3) | 195 |
Hyperparameters of the model
| Parameter | Value |
|---|---|
| Batch size | 32 |
| Number of epochs | 200 |
| Optimizer | Adam |
| Optimizer parameters | lr = 0.00001, beta1 = 0.9, verbose = 1, factor = 0.05 |
Comparison of the proposed model with other models used in the study
| Model | Number of weighted layers | Total parameters |
|---|---|---|
| ResNet18 | 18 | 11,511,784 |
| GoogLeNet | 22 | 6,797,700 |
| AlexNet | 8 | 62,378,344 |
| VGG16 | 16 | 138,423,208 |
| DenseNet161 | 161 | 28,681,000 |
| Proposed model | 8 | 130,176 |
FIGURE 6Confusion matrix
Confusion matrices of the models
|
|
FIGURE 7t‐SNE feature representation
Performance metrics
| Metric | Formula | Explanation |
|---|---|---|
| Accuracy | (TP + TN)/(TP + TN + FP + FN) | Refers to the overall success |
| Precision | TP/(TP + FP) | How accurate the positive predictions are |
| Recall | TP/(TP + FN) | Coverage of true positive samples |
| F1‐score | 2 * (Precision * Recall)/(Precision + Recall) | The harmonic mean of precision and recall |
Performance scores of models
| Model | Accuracy | Precision | Recall | F1‐score |
|---|---|---|---|---|
| ResNet18 | 0.9925 | 0.9713 | 0.9772 | 0.9742 |
| DenseNet161 | 0.9894 | 0.9595 | 0.9708 | 0.9650 |
| Custom CNN | 0.9881 | 0.9565 | 0.9649 | 0.9606 |
| GoogLeNet | 0.9855 | 0.9428 | 0.9656 | 0.9535 |
| VGG16 | 0.9831 | 0.9352 | 0.9656 | 0.9493 |
| AlexNet | 0.9703 | 0.8873 | 0.9261 | 0.9014 |
Comparison of the results obtained in this study with the literature
| Reference | Number of samples | Method | The highest Acc (%) |
|---|---|---|---|
| Chouhan et al. | 5229 | AlexNet | 92.86 |
| ResNet18 | 94.23 | ||
| Chen et al. | 35,355 | ResNet50 | 95.24 |
| Liang and Zheng | 5856 | Custom CNN | 90.5 |
| Wang et al. | 13,975 | Custom CNN | 92.6 |
| Hemdan et al. | 50 | VGG19 | 90 |
| Sethy and Behera | 381 | ResNet50 + SVM | 98.66 |
| Özkaya et al. | 300 | Custom CNN | 95.60 |
| Nour et al. | 2905 | Custom CNN + SVM | 98.97 |
| Apostolopoulos | 1427 | Vgg19 | 93.48 |
| Ghoshal and Tucker | 70 | Bayesian CNN | 92.9 |
| Uçar and Korkmaz | 5310 | SqueezeNet + Bayesian optimization | 98.26 |
| Ismael and Şengür | 380 | ResNet50 + Linear kernel SVM | 94.7 |
| Developed CNN | 91.6 | ||
| Wang et al. | 1065 | InceptionV3 | 89.5 |
| Jain et al. | 6432 | Xception | 97.97 |
| Nayak et al. | 500 | ResNet34 | 98.33 |
| Turkoglu | 6092 | Alexnet + Relief + SVM | 99.18 |
| Narin et al. | 3141 | ResNet50 | 96.1 |
| Khan et al. | 2800 | Developed CNN | 95.1 |
| Shankar and Perumal | 273 | InceptionV3 | 94.08 |
| Hammoudi et al. | 5863 | DenseNet169 | 95.72 |
| Pham | 1314 | AlexNet | 96.46 |
| Aksoy and Salman | 1019 | CapsNet | 98.02 |
| Ravi et al. | 17,599 | EfficientNet | 99 |
| This study | 15,153 | ResNet18 |
|
| DenseNet161 | 98.94 | ||
| Custom CNN | 98.81 | ||
| GoogLeNet | 98.55 | ||
| VGG16 | 98.31 | ||
| AlexNet | 97.03 |
FIGURE 8Test application
Test results
| Expert | Number of correct predictions | Number of wrong predictions | Success rate (%) | Confusions |
|---|---|---|---|---|
| 1 | 149 | 1 | 99.33 | COVID‐19—other viral (1) |
| 2 | 150 | 0 | 100 | ‐ |
| 3 | 148 | 2 | 98.66 | COVID‐19—other viral (2) |
| 4 | 146 | 4 | 97.33 |
COVID‐19—other viral (3) other viral—COVID‐19 (1) |
| 5 | 150 | 0 | 100 | ‐ |
| 6 | 148 | 2 | 98.66 | COVID‐19—other viral (1) other viral—normal (1) |
| 7 | 149 | 1 | 99.33 | Other viral—normal (1) |
| 8 | 147 | 3 | 98 | COVID‐19—other viral (3) |
| 9 | 149 | 1 | 99.33 | Other viral—COVID‐19 (1) |
| 10 | 150 | 0 | 100 | ‐ |
| 1486 | 14 | 99.06 |