| Literature DB >> 34345248 |
SeyyedMohammad JavadiMoghaddam1, Hossain Gholamalinejad2.
Abstract
The novel Coronavirus named COVID-19 that World Health Organization (WHO) announced as a pandemic rapidly spread worldwide. Fast diagnosis of the virus infection is critical to prevent further spread of the virus, help identify the infected population, and cure the patients. Due to the increasing rate of infection and the limitations of the diagnosis kit, auxiliary detection tools are needed. Recent studies show that a deep learning model that comes up with the salient information of CT images can aid in the COVID-19 diagnosis. This study proposes a novel deep learning structure that the pooling layer of this model is a combination of pooling and the Squeeze Excitation Block (SE-block) layer. The proposed model uses Batch Normalization and Mish Function to optimize convergence time and performance of COVID-19 diagnosis. A dataset of two public hospitals was used to evaluate the proposed model. Moreover, it was compared to some different popular deep neural networks (DNN). The results expressed an accuracy of 99.03 with a recognition time of test mode of 0.069 ms in graphics processing unit (GPU). Furthermore, the best network results in classification metrics parameters and real-time applications belong to the proposed model.Entities:
Keywords: Batch normalization; COVID-19 detection method; Deep learning model; Disease diagnosis; Mish function
Year: 2021 PMID: 34345248 PMCID: PMC8318781 DOI: 10.1016/j.bspc.2021.102987
Source DB: PubMed Journal: Biomed Signal Process Control ISSN: 1746-8094 Impact factor: 3.880
Fig. 1A Sample image from covid-19 CT dataset.
Fig. 2Proposed model architecture.
Configuration of the proposed network.
| layer | Output size | Kernel | Stride |
|---|---|---|---|
| Image input | 256×256×3 | – | |
| Conv | 256×256×32 | 5×5×3×32 | 1 |
| Mish + Pool + BN + Drop out (0.5) + SEBlock | 128×128×32 | – | 2 |
| Conv | 128×128×32 | 5×5×3×32 | 1 |
| Mish + Pool + BN + Drop out (0.5) + SEBlock | 64×64×32 | – | 2 |
| Conv | 64×64×64 | 5×5×3×64 | 1 |
| Mish + Pool + BN + Drop out (0.5) + SEBlock | 32×32×64 | – | 2 |
| Conv | 32×32×128 | 5×5×3×128 | 1 |
| Mish + Pool + BN + Drop out (0.5) + SEBlock | 16×16×128 | – | 2 |
| Conv | 16×16×256 | 5×5×3×256 | 1 |
| Fully connected | 65,536×256 | – | – |
| Fully connected | 256× (number of classes) | – | – |
| SoftMax | (number of classes) ×1 | – | – |
Training parameters.
| Parameter | Value |
|---|---|
| Batch size | 64 |
| Epochs | 50 |
| Momentum rate | 0.9 |
| Learning rate | 0.01 |
| Weight decay | 1e-3 |
| Epsilon | 1e-10 |
| Sampler | Weighted random sampler |
Fig. 3Block diagram of proposed network.
Metric results of the proposed network.
| Test Cohen kappa score | Test mean precision | Test mean recall | loss | Test accuracy | optimizer |
|---|---|---|---|---|---|
| 95.33 | 98.06 | 95.48 | 0.1471 | 97.15 | SGDM |
| 96.68 | 98.42 | 96.86 | 0.0869 | 97.97 | GC-SGDM |
| 95.17 | 97.98 | 95.59 | 0.0958 | 97.05 | Adam |
| 95.16 | 98.10 | 95.73 | 0.0715 | 97.06 | GC-Adam |
| 93.74 | 96.99 | 94.04 | 0.0993 | 96.19 | NAdam |
| 72.22 | 84.03 | 80.99 | 0.3780 | 82.26 | GC-NAdam |
| 98.43 | 98.71 | 98.91 | 0.0338 | 99.03 | RAdam |
| 96.27 | 98.18 | 97.02 | 0.0790 | 97.71 | GC-RAdam |
Fig. 4Comparison of max pooling and the proposed pooling layers.
Comparison of the proposed method with some popular DNNs. Prediction times do not include image loading.
| Model | Number of Parameters | Predictiontime GPU | Accuracy | Loss | Cohen kappa score | Mean precision | mean recall | |
|---|---|---|---|---|---|---|---|---|
| Popular DNNs | VGG11 + BN | 128,792,325 | 5.186 ms | 92.73 | 0.9402 | 87.87 | 95.35 | 89.42 |
| ResNet18 | 11,179,077 | 4.262 ms | 93.39 | 0.9594 | 88.98 | 95.79 | 89.49 | |
| ResNet50 | 23,518,277 | 5.554 ms | 95.98 | 0.5697 | 93.40 | 97.01 | 93.98 | |
| Inception-v3 | 24,351,719 | 6.844 ms | 98.11 | 0.1604 | 96.94 | 98.49 | 97.61 | |
| Proposed network | WCNN4 | 4,610,531 | 0.069 ms | 99.03 | 0.0338 | 98.43 | 98.71 | 98.91 |
Fig. 5Confusion matrix of Inception network.
Fig. 6Confusion matrix of proposed network.
Fig. 7Accuracy in the train phase for different DNNs.
Fig. 8Loss in train phase for different DNNs.
Fig. 9Accuracy in the train phase with test data for different DNNs.
Fig. 10Loss in train phase with test data for different DNNs.