| Literature DB >> 36034678 |
Lobna M AbouEl-Magd1,2, Ashraf Darwish3,2, Vaclav Snasel4, Aboul Ella Hassanien5,2.
Abstract
Coronavirus disease (COVID-19) is rapidly spreading worldwide. Recent studies show that radiological images contain accurate data for detecting the coronavirus. This paper proposes a pre-trained convolutional neural network (VGG16) with Capsule Neural Networks (CapsNet) to detect COVID-19 with unbalanced data sets. The CapsNet is proposed due to its ability to define features such as perspective, orientation, and size. Synthetic Minority Over-sampling Technique (SMOTE) was employed to ensure that new samples were generated close to the sample center, avoiding the production of outliers or changes in data distribution. As the results may change by changing capsule network parameters (Capsule dimensionality and routing number), the Gaussian optimization method has been used to optimize these parameters. Four experiments have been done, (1) CapsNet with the unbalanced data sets, (2) CapsNet with balanced data sets based on class weight, (3) CapsNet with balanced data sets based on SMOTE, and (4) CapsNet hyperparameters optimization with balanced data sets based on SMOTE. The performance has improved and achieved an accuracy rate of 96.58% and an F1- score of 97.08%, a competitive optimized model compared to other related models.Entities:
Keywords: COVID-19; Capsule Neural Networks; Convolution neural networks; Coronavirus; Gaussian optimization method; VGG16
Year: 2022 PMID: 36034678 PMCID: PMC9397163 DOI: 10.1007/s10586-022-03703-2
Source DB: PubMed Journal: Cluster Comput ISSN: 1386-7857 Impact factor: 2.303
Summary and analysis of the related works
| Paper | Method | Classes | Performance |
|---|---|---|---|
| Rodolf M.Pereira et al. [ | Use different feature extraction algorithms | Multiclass | The multiclass F1-score was 0.65, and the hierarchical classification F1-score was 0.89 |
| M. M. Rahaman et al. [ | VGG19 | Multiclass | Accuracy was 89.3%, and F1 score was 0.90 |
| Asmaa Abbas et al.[ | Deep CNN | Binary class | Accuracy was93.1% |
| Tej Bahadur Chandra et al. [ | Used different feature extraction techniques and used binary gray wolf optimization for feature selection. Also, used voting-based classifier ensemble is used | Distinguishes between normal and abnormal chest images and distinguishes between pneumonia, and the Covid-19 chest | Phase (I) gave 98.062% accuracy and 98.55 for the F1-score, and phase II gave 91.32% accuracy and 91.73 for F1 score The majority vote-based classifier ensemble gave an overall accuracy of 93.41% |
| O. M. Elzeki et al. [ | proposed a network architecture called CXR COVID | Multiclass | - Accuracy was 96.7% - Accuracy was 93.070% |
| M.Nour et al. [ | Suggested model based on CNN | Multiclass | Accuracy was 98.97% and an F1-score was 96.72% |
| Tulin Ozturk et al.[ | proposed the DarkNet model | binary classification multiclass classification | The binary class accuracy was 98.88%, while the multiclass accuracy was 87.02% |
| Dalia [ | DenseNet121 | The gravitational search algorithm is used to determine the best values for the hyperparameters of the DenseNet121 architecture | Accuracy 95% |
Fig. 1Capsule network architecture
Fig. 2The architecture of the pre-trained VGG16
Fig. 3The proposed COVID-19 prediction model using Optimized CapsNet
Fig. 4Visualization of VGG16 output
Fig. 5CapsNet visualization. a COVID-19. b Viral pneumonia
Fig. 6Sample images from the dataset
Parameter setting of all experiments
| Experiments | Capsule dim | Routing# | Epochs# | |
|---|---|---|---|---|
| CapsNet with the unbalanced data sets | 10 | 5 | 10 | |
| CapsNet with balanced data sets based on class weight | 10 | 5 | 20 | |
| CapsNet with balanced data sets based on SMOTE | 10 | 5 | 20 | |
| CapsNet hyperparameters optimization with balanced data sets based on SMOTE | 8 | 2 | 10 | |
Fig. 7Performance of the model with the unbalanced data sets: a Validation accuracy b model performance during the training process
Fig. 8Performance of the model with balanced data sets based on class weight: a Validation accuracy b model performance during the training process
Fig. 9Performance of the model after using SMOTE: a Validation accuracy b model performance during the training process
Iteration based CapsNet optimzation results
| iteration # | Routing | dim of capsules | Accuracy % |
|---|---|---|---|
| 1 | 4 | 10 | 90.27 |
| 2 | 3 | 10 | 87.55 |
| 3 | 3 | 9 | 89.31 |
| 4 | 4 | 11 | 89.31 |
| 5 | 2 | 4 | 91.4 |
| 6 | 2 | 8 | 93.14 |
| 7 | 1 | 13 | 92.82 |
| 8 | 2 | 10 | 88.99 |
| 9 | 2 | 12 | 87.03 |
| 10 | 4 | 16 | 93 |
| 11 | 4 | 5 | 88.7 |
Fig. 10Performance of the model after capsule hyperparameters optimization: a Validation accuracy b model performance during the training process
Experimental scenarios comparative results
| Acc (%) | P | R | F1 score | |
|---|---|---|---|---|
| CapsNet_VGG16 with imbalance data set | 89.93 | 0.8379 | 0.9090 | 0.872 |
| CapNet_VGG16 with balanced data by SMOTE | 96.73 | 0.9718 | 0.9547 | 0.9631 |
CapsNet_VGG16 with balanced data by class weight (before optimization) | 94.46 | 0.9525 | 0.9377 | 0.9400 |
CapsNet_VGG16 with balanced data by SMOTE (After optimization) | 96.58 | 0.9652 | 0.9765 | 0.9708 |
Comparison with related work
| References | Dataset size | Class # | Performance | |
|---|---|---|---|---|
| Accuracy % | F1-score | |||
| [ | 1144 | Multi class | – | 89 |
| [ | 860 | Binary class | 89.3 | 90 |
| [ | 196 | Multi class | 93.1 | – |
| [ | 582 | Binary class | 98.061 | 98.551 |
| Multi class | 91.329 | 91.73 | ||
| [ | 50 | Binary class | 92.85 | – |
| 455 | Binary class | 96.7 | – | |
| 603 | Multiclass | 93.07 | – | |
| [ | 2905 | Multi class | 98.97 | 96.7 |
| [ | Not mentioned | Binary class | 98.08 | 96.51 |
| Multi class | 87.02 | 87.02 | ||
| The proposed model | 2905 | Multi class | 96.58 | 97.08 |
Running time on the training and testing phase
| Training time (Epochs number) | Testing time | |||
|---|---|---|---|---|
| 10 teration | 15 iteration | 10 iteration | 5 iteration | 5.61173 min |
| 8.423 h | 6.32222 h | 4.22388 h | 2.119166 h | |