| Literature DB >> 33994846 |
Chao Wu1, Mohammad Khishe2, Mokhtar Mohammadi3, Sarkhel H Taher Karim4, Tarik A Rashid5.
Abstract
The COVID19 pandemic globally and significantly has affected the life and health of many communities. The early detection of infected patients is effective in fighting COVID19. Using radiology (X-Ray) images is, perhaps, the fastest way to diagnose the patients. Thereby, deep Convolutional Neural Networks (CNNs) can be considered as applicable tools to diagnose COVID19 positive cases. Due to the complicated architecture of a deep CNN, its real-time training and testing become a challenging problem. This paper proposes using the Extreme Learning Machine (ELM) instead of the last fully connected layer to address this deficiency. However, the parameters' stochastic tuning of ELM's supervised section causes the final model unreliability. Therefore, to cope with this problem and maintain network reliability, the sine-cosine algorithm was utilized to tune the ELM's parameters. The designed network is then benchmarked on the COVID-Xray-5k dataset, and the results are verified by a comparative study with canonical deep CNN, ELM optimized by cuckoo search, ELM optimized by genetic algorithm, and ELM optimized by whale optimization algorithm. The proposed approach outperforms comparative benchmarks with a final accuracy of 98.83% on the COVID-Xray-5k dataset, leading to a relative error reduction of 2.33% compared to a canonical deep CNN. Even more critical, the designed network's training time is only 0.9421 ms and the overall detection test time for 3100 images is 2.721 s.Entities:
Keywords: COVID19; Chest X-ray images; Deep convolutional neural networks; Extreme learning machine; Sine–cosine algorithm
Year: 2021 PMID: 33994846 PMCID: PMC8107782 DOI: 10.1007/s00500-021-05839-6
Source DB: PubMed Journal: Soft comput ISSN: 1432-7643 Impact factor: 3.732
Fig. 1The architecture of LeNet-5 deep CNN
Fig. 2A single-hidden layer neural network
Fig. 3The effect of sine and cosine functions on Eqs. (8) and (9)
Fig. 4Changes in sinus and cosine functions in a specified interval
Fig. 5Changes in the sinus and cosine functions within the range of [−2, 2] causes to get closer or more distant from the desired response
Fig. 6Reduction in the range of sine and cosine functions during iterations
The categories of images per class in the COVID dataset
| Category | COVID-19 | Normal |
|---|---|---|
| Training Set | 420 (84 before augmentation) | 2000 |
| Test Set | 100 | 3000 |
Fig. 7Six stochastic sample images from the COVID-X-ray-5k dataset
Fig. 8The conventional vs. proposed architecture
Fig. 9Assigning the deep CNN’s parameters as the candid solution (searching agents) of SCA
Fig. 10The pseudo-code for DCELM-SCA model
Fig. 11The flowchart of the designed model
The parameters of benchmark algorithms
| Algorithm | Parameters | Values |
|---|---|---|
| GA | Cross-over probability | 0.7 |
| Mutation probability | 0.1 | |
| Population size | 50 | |
| CS | Discovery rate of alien eggs | 0.25 |
| Population size | 50 | |
| WOA | Linearly decreased from 2 to 0 | |
| Population size | 50 | |
| SCA | 2 | |
| Random in the range of [0, 1] | ||
| Population size | 50 |
Fig. 12The EPG produced by in_6c_2p_12c_2p structure
Fig. 13The EPG produced by in_8c_2p_16c_2p structure
Fig. 14The confusion matrix for in_6c_2p_12c_2p model
Fig. 15The confusion matrix for in_8c_2p_16c_2p model
Specificity and sensitivity rates of benchmark models for various threshold values
| Model | Threshold | Sensitivity (%) | Specificity (%) | |
|---|---|---|---|---|
| deep CNN | 0.1 | 98.22 ± 0.002 | 84.47 ± 0.003 | 0.0047 |
| 0.2 | 95.35 ± 0.002 | 85.73 ± 0.002 | 0.0025 | |
| 0.3 | 90.56 ± 0.005 | 87.42 ± 0.002 | 0.0047 | |
| 0.4 | 84.54 ± 0.006 | 90.82 ± 0.002 | 0.0085 | |
| DCELM | 0.1 | 98.11 ± 0.052 | 83.37 ± 0.082 | 0.041 |
| 0.2 | 94.56 ± 0.056 | 86.21 ± 0.022 | 0.0056 | |
| 0.3 | 89.96 ± 0.085 | 88.12 ± 0.013 | 0.0056 | |
| 0.4 | 83.22 ± 0.101 | 89.52 ± 0.011 | 3.12E−06 | |
| DCELM-GA | 0.1 | 98.33 ± 0.002 | 92.26 ± 0.002 | 0.0005 |
| 0.2 | 97.21 ± 0.003 | 93.85 ± 0.002 | 0.002 | |
| 0.3 | 92.36 ± 0.005 | 94.85 ± 0.001 | 1.11E−11 | |
| 0.4 | 89.24 ± 0.005 | 96.85 ± 0.001 | 0.0004 | |
| DCELM-CS | 0.1 | 99.23 ± 0.001 | 89.91 ± 0.002 | 0.0012 |
| 0.2 | 97.63 ± 0.001 | 92.85 ± 0.002 | 0.0032 | |
| 0.3 | 95.32 ± 0.002 | 96.33 ± 0.001 | 2.79E−06 | |
| 0.4 | 91.11 ± 0.002 | 97.33 ± 0.001 | 0.003 | |
| DCELM-WOA | 0.1 | 99.01 ± 0.002 | 85.12 ± 0.004 | |
| 0.2 | 96.65 ± 0.003 | 92.98 ± 0.004 | 0.041 | |
| 0.3 | 91.21 ± 0.003 | 96.60 ± 0.003 | 0.045 | |
| 0.4 | 80.32 ± 0.004 | 97.90 ± 0.002 | 0.025 | |
| DCELM-SCA | 0.1 | 100 ± 0.000 | 84.34 ± 0.002 | N/A |
| 0.2 | 98.12 ± 0.001 | 93.32 ± 0.001 | N/A | |
| 0.3 | 97.56 ± 0.001 | 95.33 ± 0.001 | N/A | |
| 0.4 | 92.99 ± 0.002 | 98.66 ± 0.001 | N/A |
The reliability analysis of sensitivity and specificity of four evolutionary benchmark DCELM and deep CNN
| Model | Sensitivity (%) | Specificity (%) |
|---|---|---|
| deep CNN | 98 ± 2.8 | 84.47 ± 1.31 |
| DCELM | 98 ± 2.8 | 83.37 ± 1.32 |
| DCELM-GA | 98 ± 2.8 | 92.26 ± 0.90 |
| DCELM-CS | 98 ± 2.8 | 91.85 ± 0.91 |
| DCELM-WOA | 98 ± 2.8 | 91.33 ± 0.91 |
| DCELM-SCA | 98 ± 2.8 | 93.22 ± 0.82 |
Fig. 16The ROC curves and precision-recall curves for DCELM-SCA and other benchmarks
The comparison of test and training time of benchmark network implemented on GPU and CPU
| Model | CPU vs. GPU | Training time | Testing time | |
|---|---|---|---|---|
| deep CNN | GPU | 10 min, 34 s | 3180 ms | 1.53E−07 |
| CPU | 6 h, 44 min, 8 s | 4 min, 30 s | 1.37E−03 | |
| DCELM | GPU | |||
| CPU | ||||
| DCELM-GA | GPU | 3645.6 ms | 3102 ms | 1.13E−03 |
| CPU | 4 min, 26.6 s | 4 min, 22 s | 1.05E−04 | |
| DCELM-CS | GPU | 2578.2 ms | 3101 ms | 1.62E−05 |
| CPU | 3 min, 9.2 s | 4 min, 27 s | 1.32E−03 | |
| DCELM-WOA | GPU | 1299.2 ms | 3015 ms | |
| CPU | 2 min, 9 s | 4 min, 21 sec | 1.45E−09 | |
| DCELM-SCA | GPU | 1287 ms | 2985 ms | |
| CPU | 2 min, 01 s | 4 min, 20 sec |
The specification of parameters
| Parameters | Level | |||
|---|---|---|---|---|
| 1 | 2 | 3 | 4 | |
| 3 | 4 | 5 | 6 | |
| 2.5 | 0.5 | 0.75 | 1 | |
| 6 | 8 | 10 | 12 | |
Results of various parameter combinations
| Test number | Parameters | Result (MSE) | ||
|---|---|---|---|---|
| Ex. #1 | 3 | 0.25 | 6 | 0.1984 |
| Ex. #2 | 3 | 0.5 | 8 | 0.1652 |
| Ex. #3 | 3 | 0.75 | 10 | 0.1655 |
| Ex. #4 | 3 | 1 | 12 | 0.0952 |
| Ex. #5 | 4 | 0.25 | 8 | 0.1821 |
| Ex. #6 | 4 | 0.5 | 6 | 0.1852 |
| Ex. #7 | 4 | 0.75 | 12 | 0.1123 |
| Ex. #8 | 4 | 1 | 10 | 0.0852 |
| Ex. #9 | 5 | 0.25 | 6 | 0.0923 |
| Ex. #10 | 5 | 0.5 | 12 | 0.0601 |
| Ex. #11 | 5 | 0.75 | 8 | 0.0532 |
| Ex. #12 | 5 | 1 | 10 | 0.0423 |
| Ex. #13 | 6 | 0.25 | 12 | 0.0977 |
| Ex. #14 | 6 | 0.5 | 10 | 0.0887 |
| Ex. #15 | 6 | 0.75 | 8 | 0.0731 |
| Ex. #16 | 6 | 1 | 6 | 0.0511 |
Fig. 17Level trends of the analyzed parameters
Fig. 18Search history, convergence curve, average fitness history, and trajectory of some functions
Benchmark functions
| Function | Dim | Range | |
|---|---|---|---|
| 30 | 0 | ||
| 30 | 0 | ||
| 30 | |||
| 30 | 0 | ||
| 30 | [− 50, 50] | 0 | |
| 2 | [− 65, 65] | 1 | |
| 2 | [− 2, 2] | 3 | |
| 10 | [− 5, 5] | 0 | |
| 10 | [− 5, 5] | 0 | |
Fig. 19ROI for positive COVID19 cases using ACM
Fig. 20ROI for Normal cases using ACM