| Literature DB >> 33738639 |
Mukul Singh1, Shrey Bansal1, Sakshi Ahuja2, Rahul Kumar Dubey3, Bijaya Ketan Panigrahi2, Nilanjan Dey4.
Abstract
The novel discovered disease coronavirus popularly known as COVID-19 is caused due to severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) and declared a pandemic by the World Health Organization (WHO). An early-stage detection of COVID-19 is crucial for the containment of the pandemic it has caused. In this study, a transfer learning-based COVID-19 screening technique is proposed. The motivation of this study is to design an automated system that can assist medical staff especially in areas where trained staff are outnumbered. The study investigates the potential of transfer learning-based models for automatically diagnosing diseases like COVID-19 to assist the medical force, especially in times of an outbreak. In the proposed work, a deep learning model, i.e., truncated VGG16 (Visual Geometry Group from Oxford) is implemented to screen COVID-19 CT scans. The VGG16 architecture is fine-tuned and used to extract features from CT scan images. Further principal component analysis (PCA) is used for feature selection. For the final classification, four different classifiers, namely deep convolutional neural network (DCNN), extreme learning machine (ELM), online sequential ELM, and bagging ensemble with support vector machine (SVM) are compared. The best performing classifier bagging ensemble with SVM within 385 ms achieved an accuracy of 95.7%, the precision of 95.8%, area under curve (AUC) of 0.958, and an F1 score of 95.3% on 208 test images. The results obtained on diverse datasets prove the superiority and robustness of the proposed work. A pre-processing technique has also been proposed for radiological data. The study further compares pre-trained CNN architectures and classification models against the proposed technique.Entities:
Keywords: COVID-19; CT scan data; Ensemble SVM; Transfer learning; VGG16
Mesh:
Year: 2021 PMID: 33738639 PMCID: PMC7972022 DOI: 10.1007/s11517-020-02299-2
Source DB: PubMed Journal: Med Biol Eng Comput ISSN: 0140-0118 Impact factor: 2.602
Summary of techniques available in literature for COVID19 screening
| Ref. | Technique | Key findings | Dataset |
|---|---|---|---|
| [ | Transfer learning on ResNet-50 (CNN model) | Achieved 10-fold cross-validation accuracy of 93.01% on 109 test images. | 413 COVID-19 (+) images and 439 images of normal or pneumonia infected patients. [ |
| [ | Resnet50 and VGG16 (deep learning) | COVID-19 positive cases and pneumonia cases of X-ray modalities are classified with an accuracy of 89.2%. | There were 135 COVID-19 cases obtained from JP Cohen [ |
| [ | Two-step transfer learning model | Two-step transfer learning pipeline based on the deep neural network framework COVID19XrayNet. The approach achieved a maximum accuracy of 91.4%. | The study uses 189 COVID-19 Chest X-Ray images (131 train and 41 test). [ |
| [ | EfficientNet | An accuracy of 93.9%, a sensitivity of 96.8%, and positivity prediction of 100% are obtained on 231 test X-ray images (COVID19-positive cases—31, pneumonia—100, and normal cases—100). | The model is trained on 13569 X-ray images (COVID19 positive cases-152, pneumonia-5421, and normal cases-7966) [ |
| [ | Pre-trained CheXNet and DenseNet | An accuracy of 90.5% and a sensitivity of 100% are achieved on 654 test X-ray images (COVID19-positive cases—30, pneumonia—390, and normal cases—234). | The model is trained on 5323 chest X-ray images (COVID19 positive cases-115, pneumonia-3867, and normal cases - 1341) [ |
| [ | Domain extension transfer learning (DETL) with gradient class activation map (Grad-CAM) | Fivefold cross-validation accuracy of 90.13% and test set accuracy of 95.3% are obtained on the proposed X-ray dataset. | A total of 305 COVID-19 X-Ray images were used in the study. [ |
| [ | ResNet, Inception, and GoogleNet | The classification of COVID-19 positive cases based on X-ray modality is done. The approach achieved 98% of accuracy with VGG19, 95% with Resnet50, and 96% with InceptionV3. | A dataset of nearly 100 subject, among them 50 x-ray images subjects were tested positive with COVID-19 and 50 x-ray images of normal subjects. [ |
| [ | ResNet18, ResNet50, SqueezeNet, and DenseNet-121 | The model achieved a specificity of 90% and sensitivity of 96.5% on testing data of 3000 chest X-rays (COVID and non-COVID patients). | The model is trained on 5000 chest x-rays dataset [ |
| [ | Joint classification and segmentation (JCS) | Classification is done with a specificity of 93% and a sensitivity of 95%. A dice score of 78.3% is obtained for the segmentation task. | JCS system is implemented on 400 COVID-19 patients (144,167 images) and 350 Non-COVID patients. [ |
| [ | Pruned efficient net-based model on chest CT scans and X-rays | Classification into two binary classes, i.e., COVID and non-COVID. The highest accuracy of 85.22% is achieved with the ResNet50 pre-trained CNN model. | The CNN based pre-trained models are trained on 596 chest CT scans. [ |
| [ | Detail-oriented capsule networks (DECAPS) +Peekaboo (patch crop and drop strategy) | An accuracy of 87.6%, recall of 91.5%, precision of 84.3%, and AUC of 96.1 are achieved for binary classification (COVID-19 and non-COVID) of chest CT scan. | Uses a total of 746 chest CT images - COVID-19 (349 images) and non-COVID-19 (347 images). [ |
| [ | Transfer learning on Xception net | For binary classification of chest CT scan of COVID and non-COVID dataset, the model achieved a sensitivity of 96.1%, the specificity of 93.4%, and AUC of 0.92. | It contains three classes as COVID-19 (+), pneumonia (+) but COVID-19 (-) with 504 images. [ |
| [ | Multi-objective differential evolution (MODE) deep learning | In comparison to authentic CNN models, the performance parameters of MODE outperforms by 2.09% of | A study of 73 patients with 205 COVID positive images. [ |
Fig. 1Self-explanatory block diagram of the proposed methodology of COVID-19 screening
The brief details of the dataset for the proposed model
| Dataset | COVID | Non-COVID | Total |
|---|---|---|---|
| D1 | 233 images (training—204, and validation—29) | 358 images (training—228, validation—33, and test—97) | 591 |
| D2 | 53 images (test—53) | 0 images | 53 |
| D3 | 58 images (test—58) | 0 images | 58 |
| Total | 344 images | 358 images | 702 |
Fig. 2Pictorial representation of various stages of the pre-processing module
Comparative study of various popular CNN architectures
| Sr. No. | CNN architecture | Accuracy on validation set (%) |
|---|---|---|
| 1 | VGG16 | 79.1 |
| 2 | VGG19 | 77.2 |
| 3 | Resnet50 | 70.8 |
| 4 | InceptionV3 | 72.2 |
| 5 | DenseNet21 | 68.5 |
Fig. 3Architecture of truncated VGG16 model
Summary of various VGG16 truncation point accuracy evaluated on the validation set with SVM as classifier
| Sr. No. | Truncation point | Accuracy on validation set (%) |
|---|---|---|
| 1 | 3 blocks | 73.6 |
| 2 | 4 blocks | 84.2 |
| 3 | Un-truncated | 79.1 |
Fig. 4Intermediate color-mapped outputs. a Layer 1. b Layer 4. c Layer 8. d Layer 14
Fig. 5Comparision of confusion matrices before and after fine-tuning of VGG16 by evaluation on the test set with bagging SVM as the classifier
Performances analysis of feature selection techniques on validation set using SVM as classifier
| Sr. No. | Feature selection technique | Validation accuracy (%) |
|---|---|---|
| 1 | PCA | 93.4 |
| 2 | Autoencoder | 89.6 |
| 3 | Variance-based selector | 87.3 |
Fig. 8Confusion matrices of the proposed methodology with different classifiers
Performance parameters of different classifiers on testing data
| Classifier | TP | TN | FP | FN | AUC | PRE | NPV | S1 | S2 | F1 | ACC |
|---|---|---|---|---|---|---|---|---|---|---|---|
| Bagging with SVM | 107 | 92 | 5 | 4 | 95.8 | 0.955 | 0.958 | 0.963 | 0.948 | 0.959 | 0.957 |
| ELM | 107 | 88 | 9 | 4 | 93.8 | 0.922 | 0.956 | 0.963 | 0.907 | 0.942 | 0.937 |
| OS-ELM | 107 | 90 | 7 | 4 | 94.9 | 0.938 | 0.957 | 0.963 | 0.927 | 0.951 | 0.947 |
| Deep CNN | 103 | 82 | 15 | 8 | 89.5 | 0.872 | 0.911 | 0.927 | 0.845 | 0.899 | 0.889 |
Fig. 6Convergence graph of accuracy vs epoch for proposed methodology (VGG16+PCA+bagging ensemble with SVM)
Fig. 7Learning curve for proposed method using 10-fold cross-validation
Fig. 9ROC characteristics curve for the proposed methodology (VGG16+PCA+bagging ensemble with SVM)
Comparative analysis of COVID-19 detection proposed methodology with techniques available in the literature on the used dataset
| Sr. No. | Techniques | Dataset | Performance evaluation |
|---|---|---|---|
| 1. | DECAPS + Peekaboo [ | Binary classification of total 746 chest CT images COVID-19 and non-COVID-19 [ | Accuracy - 87.6%, AUC- 0.961, and precision - 84.3%. |
| 2. | Resnet50 and VGG16 [ | Total 102 X-ray images of COVID-19-positive and -pneumonia patients [ | Overall accuracy achieved is 89.2%. |
| 3. | AI methods (JCS and DenseNet169) [ | Binary classification of CT scan data into COVID (349 images) and non-COVID (463 CT images) [ | Accuracy - 0.83, AUC - 0.95, and F1 - 0.85. |
| 4. | Proposed methodology VGG16+PCA+Bagging Ensemble with SVM | Binary classification (COVID-19 and non-COVID-19) using 702 CT scan images (344 COVID-19 images and 358 non-COVID images) [ | Prediction time is 385ms, Accuracy - 95.7%, Precision - 95.8%, AUC - 0.958, and F1 score - 95.3%. |