Literature DB >> 33738639

Transfer learning-based ensemble support vector machine model for automated COVID-19 detection using lung computerized tomography scan data.

Mukul Singh1, Shrey Bansal1, Sakshi Ahuja2, Rahul Kumar Dubey3, Bijaya Ketan Panigrahi2, Nilanjan Dey4.   

Abstract

The novel discovered disease coronavirus popularly known as COVID-19 is caused due to severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) and declared a pandemic by the World Health Organization (WHO). An early-stage detection of COVID-19 is crucial for the containment of the pandemic it has caused. In this study, a transfer learning-based COVID-19 screening technique is proposed. The motivation of this study is to design an automated system that can assist medical staff especially in areas where trained staff are outnumbered. The study investigates the potential of transfer learning-based models for automatically diagnosing diseases like COVID-19 to assist the medical force, especially in times of an outbreak. In the proposed work, a deep learning model, i.e., truncated VGG16 (Visual Geometry Group from Oxford) is implemented to screen COVID-19 CT scans. The VGG16 architecture is fine-tuned and used to extract features from CT scan images. Further principal component analysis (PCA) is used for feature selection. For the final classification, four different classifiers, namely deep convolutional neural network (DCNN), extreme learning machine (ELM), online sequential ELM, and bagging ensemble with support vector machine (SVM) are compared. The best performing classifier bagging ensemble with SVM within 385 ms achieved an accuracy of 95.7%, the precision of 95.8%, area under curve (AUC) of 0.958, and an F1 score of 95.3% on 208 test images. The results obtained on diverse datasets prove the superiority and robustness of the proposed work. A pre-processing technique has also been proposed for radiological data. The study further compares pre-trained CNN architectures and classification models against the proposed technique.

Entities:  

Keywords:  COVID-19; CT scan data; Ensemble SVM; Transfer learning; VGG16

Mesh:

Year:  2021        PMID: 33738639      PMCID: PMC7972022          DOI: 10.1007/s11517-020-02299-2

Source DB:  PubMed          Journal:  Med Biol Eng Comput        ISSN: 0140-0118            Impact factor:   2.602


Introduction

A chest infection is a kind of infection that affects the proper functioning of the lungs (both larger and smaller airways) [1]. The severity of a lung infection depends on several factors like causes of lung infection (virus or bacteria) and the overall health of the infected person. The most common lung infections are pneumonia, chronic obstructive pulmonary disease (COPD), asthma, bronchitis, and lung cancer. Coronavirus disease popularly known as COVID-19 is a kind of lung infection disease. It is caused due to the novel discovered virus know as severe acute respiratory syndrome coronavirus 2 (SARS- CoV-2). Coronaviruses are a family of viruses that are known to cause diseases like a common cold, severe acute respiratory syndrome (SARS), and Middle East respiratory syndrome (MERS)[2]. The coronavirus disease is first discovered in Wuhan, China, in December 2019. The unprecedented rise in COVID-19 cases is impacting the worldwide economy and declared a pandemic by the World Health Organization [3]. On 22 May 2020, a total 5,207,918 patients are infected with COVID-19, and 334,848 deaths are reported across 215 countries [4]. To control the spread of the COVID-19 virus, its accurate detection and treatment are required. Reverse transcriptase polymerase chain reaction (RT- PCR) is the standard diagnostic test for COVID-19 [5]. The high popularity of PCR is due to its high selectivity and sensitivity, i.e., over 90%. The limitations of the COVID-19 testing with PCR technique are (a) very time consuming, (b) expensive, and (c) shortage of kits due to long production time [6]. Considering the alarming rates of spread of COVID-19, a faster and cheaper testing mechanism is required to tackle this outbreak. The need for a faster screening technique to control the COVID outbreak has also been studied by the authors in [7, 8]. Researchers have found that radiological analysis like X-rays and chest CT (computed tomography) scans have high accuracy in COVID-19 diagnosis and can be an effective tool for large scale screening. A high correlation between RT-PCR and radiological results for COVID-19 diagnosis is established in [9]. Also, COVID-19 infection is identified through ground-glass opacity patches (GGO) in radiographic scans of patients. This encouraged the development of a faster and cheaper COVID-19 screening mechanism using a radiological approach. Also, deep learning is playing a critical role in medical image analysis which motivates its use in screening of COVID-19. The growing role of deep learning has also been analyzed by the study of AI-based COVID classification techniques [10] where authors find that deep-learning-based techniques can provide very promising results for COVID classification. The details of the techniques available in the literature for COVID-19 diagnosis is put forth in Table 1.
Table 1

Summary of techniques available in literature for COVID19 screening

Ref.TechniqueKey findingsDataset
[11]Transfer learning on ResNet-50 (CNN model)Achieved 10-fold cross-validation accuracy of 93.01% on 109 test images.413 COVID-19 (+) images and 439 images of normal or pneumonia infected patients. [12]
[13]Resnet50 and VGG16 (deep learning)COVID-19 positive cases and pneumonia cases of X-ray modalities are classified with an accuracy of 89.2%.There were 135 COVID-19 cases obtained from JP Cohen [14], radiopedia and SIRM [15].
[16]Two-step transfer learning modelTwo-step transfer learning pipeline based on the deep neural network framework COVID19XrayNet. The approach achieved a maximum accuracy of 91.4%.The study uses 189 COVID-19 Chest X-Ray images (131 train and 41 test). [14]
[17]EfficientNetAn accuracy of 93.9%, a sensitivity of 96.8%, and positivity prediction of 100% are obtained on 231 test X-ray images (COVID19-positive cases—31, pneumonia—100, and normal cases—100).The model is trained on 13569 X-ray images (COVID19 positive cases-152, pneumonia-5421, and normal cases-7966) [18]
[19]Pre-trained CheXNet and DenseNetAn accuracy of 90.5% and a sensitivity of 100% are achieved on 654 test X-ray images (COVID19-positive cases—30, pneumonia—390, and normal cases—234).The model is trained on 5323 chest X-ray images (COVID19 positive cases-115, pneumonia-3867, and normal cases - 1341) [14].
[20]Domain extension transfer learning (DETL) with gradient class activation map (Grad-CAM)Fivefold cross-validation accuracy of 90.13% and test set accuracy of 95.3% are obtained on the proposed X-ray dataset.A total of 305 COVID-19 X-Ray images were used in the study. [14]
[21]ResNet, Inception, and GoogleNetThe classification of COVID-19 positive cases based on X-ray modality is done. The approach achieved 98% of accuracy with VGG19, 95% with Resnet50, and 96% with InceptionV3.A dataset of nearly 100 subject, among them 50 x-ray images subjects were tested positive with COVID-19 and 50 x-ray images of normal subjects. [14, 15]
[22]ResNet18, ResNet50, SqueezeNet, and DenseNet-121The model achieved a specificity of 90% and sensitivity of 96.5% on testing data of 3000 chest X-rays (COVID and non-COVID patients).The model is trained on 5000 chest x-rays dataset [14]
[23]Joint classification and segmentation (JCS)Classification is done with a specificity of 93% and a sensitivity of 95%. A dice score of 78.3% is obtained for the segmentation task.JCS system is implemented on 400 COVID-19 patients (144,167 images) and 350 Non-COVID patients. [14, 24, 25]
[26]Pruned efficient net-based model on chest CT scans and X-raysClassification into two binary classes, i.e., COVID and non-COVID. The highest accuracy of 85.22% is achieved with the ResNet50 pre-trained CNN model.The CNN based pre-trained models are trained on 596 chest CT scans. [14, 15]
[27]Detail-oriented capsule networks (DECAPS) +Peekaboo (patch crop and drop strategy)An accuracy of 87.6%, recall of 91.5%, precision of 84.3%, and AUC of 96.1 are achieved for binary classification (COVID-19 and non-COVID) of chest CT scan.Uses a total of 746 chest CT images - COVID-19 (349 images) and non-COVID-19 (347 images). [15]
[28]Transfer learning on Xception netFor binary classification of chest CT scan of COVID and non-COVID dataset, the model achieved a sensitivity of 96.1%, the specificity of 93.4%, and AUC of 0.92.It contains three classes as COVID-19 (+), pneumonia (+) but COVID-19 (-) with 504 images. [25]
[29]Multi-objective differential evolution (MODE) deep learningIn comparison to authentic CNN models, the performance parameters of MODE outperforms by 2.09% of F-measure, 1.82% of sensitivity, 1.68% of specificity, and 1.927% of Kappa statistics.A study of 73 patients with 205 COVID positive images. [30]
From the detailed analysis of the state of the art of COVID-19 diagnosis field, it can be inferred that chest radiography (X-rays and CT scan) is the best alternative for COVID-19 detection in comparison to the RT-PCR test kits [31]. However, CT scan modality seems to be most efficient in comparison to chest X-ray due to the following reasons: (a) CT scan gives a detailed 3-dimensional view of the diagnosed organ whereas X-rays give a 2-D view, (b) the CT scan does not overlap the organ, whereas in X-rays ribs overlap the lungs and heart. Due to the high precision of a CT-scan based screening system, a deep learning-based 3 step model is proposed which consists of a transfer-learning-based feature extractor, a feature selector, and a feature classifier. In the proposed work, a truncated VGG16 architecture is proposed for extracting features. The last two blocks of the truncated architecture are fine-tuned with differential learning rates. PCA is applied to the features extracted by CNN. For the classification task, four different classifier models are compared. The study addresses various issues with the current COVID-19 datasets and proposes various techniques to overcome these. Transfer learning capabilities of various models have been demonstrated and compared. Furthermore, various techniques like truncation and differential learning rates are proposed to increase robustness. The effect of various feature selection techniques has also been studied. Finally, bagging SVM is chosen for classification after a comparative study of popular classifiers. The proposed model within 385 ms achieved an accuracy of 95.7%, the precision of 95.8%, area under curve (AUC) of 0.958, and an F1 score of 95.3% on the 208 test images. The results obtained on diverse datasets prove the superiority and robustness of the proposed work. The rest of the paper is organized as follows: Section 2 illustrates the proposed methodology; Section 2.7 put-forth the details of different classifiers. Section 3 gives the details of results and discussion. Then Section 5 concludes the proposed work. Summary of techniques available in literature for COVID19 screening

Proposed methodology

The chest CT scans of COVID-19 patients contain patches of ground glass opacity (GGO); thus, a multi-dimensional feature extractor is required for screening [32]. In the proposed work, the VGG16 architecture is fine-tuned and used to extract features from lung CT scan images. Since the size of the COVID-19 dataset is very small, a truncated version of the VGG16 architecture is used. PCA is used to reduce the dimensionality of the features obtained from truncated VGG-16. The final classification is performed using four different classifiers. The self-explanatory block diagram of the proposed methodology for COVID-19 classification is shown in Fig. 1.
Fig. 1

Self-explanatory block diagram of the proposed methodology of COVID-19 screening

Self-explanatory block diagram of the proposed methodology of COVID-19 screening

Training data

In the proposed work, the dataset is collected from three different sources to ensure the robustness of the model. The brief details of datasets used are: Dataset 1 (D1)- A CT scan dataset of 617 COVID and non-COVID images, compiled by Zhao et. al [14] Dataset 2 (D2)- COVID-19 image data collection (53 COVID CT scans): Joseph Paul Cohen, Paul Morrison, and Lan Dao [24] Dataset 3 (D3)- Italian society of medical and interventional research (60 COVID-19 CT scans) [15] The final split of the data is summarized below: Training : 432 images (204 COVID and 228 non-COVID before augmentation Validation : 62 images (29 COVID and 33 non-COVID) Test : 208 images (111 COVID and 97 non-COVID) Some images in these datasets had markings and other non-removable artifacts and hence had to be dropped. Table 2 puts forth the details of CT scan images available in D1, D2, and D3 along-with the details of training, validation, and test set used. The minimum, average, and maximum height are 153, 491, and 1853. The minimum, average, and maximum width are 124, 383, and 1485. These images are from 216 patient cases. For patients labeled with positive, 169 of them have age information and 137 of them have gender information. The images are from multiple sources including hospital donations, pre-prints, and reports released by the ISMIR [15]. The exact patient and image details can be found at [14, 15, 24].
Table 2

The brief details of the dataset for the proposed model

DatasetCOVIDNon-COVIDTotal
D1233 images (training—204, and validation—29)358 images (training—228, validation—33, and test—97)591
D253 images (test—53)0 images53
D358 images (test—58)0 images58
Total344 images358 images702

Pre-processing module

As the input images are of different sizes, thus all the input images are resized to 112 × 112 × 3 to maintain the uniformity. CT scans have artifacts like beam hardening, noise, and scatter, which reduce the accuracy of the model. To overcome this, first, a median filter has been applied. Median filtering is a widely used nonlinear method used to remove noise from images while preserving edges. The median filter operates by replacing each pixel value with the median value of neighboring pixels. 5 × 5 is a popular median filter kernel choice for biomedical images and has hence been chosen [33]. Finally, morphological close transformation is applied to the image. A morphological close operation is a dilation operation followed by an erosion operation. It removes holes and any remaining salt and pepper noise from the images. It has been shown to be highly effective on binary and gray-scale images [34]. Since the images are of different scales and have labelings and markings around corners, an adaptive ROI selector has been applied to the images. The filter first centers and straightens the image, then applies an elliptical mask on the images such that the non-lung parts get cropped out. The ellipse is fitted to the image using the abrupt pixel-value changes as we move close to the rib-cage bones (black to white color change). Furthermore, all the masked images are manually checked to ensure that no image is over or under-cropped. Such images are manually fixed. Figure 2 shows the pictorial representation of the various pre-processing module used in the study.
Fig. 2

Pictorial representation of various stages of the pre-processing module

Pictorial representation of various stages of the pre-processing module The brief details of the dataset for the proposed model

Image augmentation

Data augm‘entation allows the model to learn a more diverse set of features and also increases the size of the dataset thereby preventing the model from overfitting. Each training image is augmented by a random affine transformation, random flip, and random changes in hue, brightness, and saturation of the image. The random affine transformation consists of shearing and rotation. The details of image augmentation parameters include (a) rotation—within range of 0 to 30∘, (b) shearing—0.2, (c) zooming—0.2, and (d) changing the brightness level—within range of 0.75 to 1.5. The augmentation parameters were chosen based on the study of the effectiveness of image augmentation techniques on deep networks [35]. The parameters chosen have also been used in [36] where authors have achieved great results with these settings on a CT scan classification problem similar to this. The training data after augmentation is (a) 612 of COVID-19 images and (b) 684 of non-COVID-19 images.

VGG-based feature extractor

Table 3 shows the result of the comparative study considering the popular CNN architectures. The accuracy reported is on the test set by adding a classification layer to the model, and it can be seen that VGG16 outperforms other deeper architectures. This is an interesting observation that deeper models are performing poorer on the current COVID-19 datasets. This is probably due to the size and quality of the currently available datasets. Due to the better performance of VGG16, it is the model of choice in this study. Similar observations have been made in [37] where authors use VGG16 to classify COVID-19 from a multi-modal input.
Table 3

Comparative study of various popular CNN architectures

Sr. No.CNN architectureAccuracy on validation set (%)
1VGG1679.1
2VGG1977.2
3Resnet5070.8
4InceptionV372.2
5DenseNet2168.5
Initially, the VGG model is trained on the ImageNet database with over 14 million images [38]. Instead of using large receptive fields, VGG16 uses very small receptive fields (3 × 3 with a stride of 1). VGG16 incorporates 1 × 1 convolution layers to make the decision function more non-linear without changing the receptive fields. Since the COVID-19 dataset is much smaller with only 591 training images (before augmentation), the high complexity of the feature set will be difficult to generalize. To prevent this, a truncated VGG16 architecture is proposed which limits the complexity of the features. The first four convolution blocks of the VGG16 architecture are used for the proposed truncated architecture as shown in Fig. 3. The truncation of the architecture reduces model complexity and number of trainable parameters, which eventually helps in reducing overfitting. This technique has also been used in Inception-net based COVID classification in [39] where the authors truncated the architecture to reduce overfitting. The truncation layer is determined by evaluating performance on the validation set with different points of truncation as detailed out in Table 4.
Fig. 3

Architecture of truncated VGG16 model

Table 4

Summary of various VGG16 truncation point accuracy evaluated on the validation set with SVM as classifier

Sr. No.Truncation pointAccuracy on validation set (%)
13 blocks73.6
24 blocks84.2
3Un-truncated79.1
Architecture of truncated VGG16 model

Transfer learning

Training a neural network from scratch requires huge amounts of data. As the COVID-19 dataset available is significantly smaller, transfer learning is applied to extract an accurate and concise feature set from the training data. This is a popular technique and has also achieved great results in [11, 13, 17, 19]. Comparative study of various popular CNN architectures In the proposed methodology, a representation learning-based approach is used. A pre-trained VGG-16 model is fine-tuned and its intermediate outputs act as a representation of raw data. This representation serves as features for the classifier module. The first four blocks of the VGG16 architecture pre-trained on ImageNet weights are used for this purpose [13, 17]. Since the Image net set is non-overlapping to the problem, the last 8 layers, i.e., the third and fourth convolution blocks are fine-tuned on the augmented CT scan training data [40]. While training these, it is desired that the fourth block adapts more to the data compared to the third block. The third block carries relatively fewer complex features that do not need to change much. Hence, a higher learning rate has been used for the fourth convolutional block compared to the third convolutional block while fine-tuning [11]. The extracted features are displayed as a color map as shown in Fig. 4. Figure 5 shows the confusion matrices of the proposed model with and without fine-tuning of the VGG16 based feature extractor.
Fig. 4

Intermediate color-mapped outputs. a Layer 1. b Layer 4. c Layer 8. d Layer 14

Fig. 5

Comparision of confusion matrices before and after fine-tuning of VGG16 by evaluation on the test set with bagging SVM as the classifier

The feature extractor module reduces the dimension of the data to 25,000 features per image for an image size of 112 × 112 × 3 pixels. However, with only 591 training examples (before augmentation), the model would still overfit the features. To prevent this, feature selection and dimensionality reduction of data are performed. Intermediate color-mapped outputs. a Layer 1. b Layer 4. c Layer 8. d Layer 14 Comparision of confusion matrices before and after fine-tuning of VGG16 by evaluation on the test set with bagging SVM as the classifier

Feature selector

Principal component analysis (PCA), autoencoders, and variance-based selectors are the most popular feature selectors for image data. [41] finds PCA to perform significantly better as a feature selector on biomedical data. PCA finds the eigenvectors of a covariance matrix with the highest eigenvalues and then uses those to project the data into a new subspace of equal or fewer dimensions. Autoencoders compress the input to a lower dimension. Variance-based methods select the features which have the highest variance over the data. PCA, autoencoder, and variance-based selector have been used to reduce the dimensionality of the feature set, and then their accuracies on the validation set are compared after classification with an SVM. Applying PCA with 95% variance representation yields 359 components. Since 95% variance is a standard value for variance-based reductions, the autoencoder and variance-based selectors were also configured to retain 95% variance of the original feature set. The results of the analysis are tabulated in Table 5. For the proposed model, PCA gives the highest accuracy because it represents the low-dimensional sample and synchronized variables. Furthermore, the extracted features from the training set are used to train the classification module to screen COVID-19 CT scans. The better performance of PCA as a feature selector has also been emphasized in [42].
Table 5

Performances analysis of feature selection techniques on validation set using SVM as classifier

Sr. No.Feature selection techniqueValidation accuracy (%)
1PCA93.4
2Autoencoder89.6
3Variance-based selector87.3

Classification

For the classification task, the required features are extracted using the truncated VGG16 model and selected using PCA. In machine-learning, no single algorithm is suitable for all problems. Thus, for achieving the highest performance, 4 different classification models are evaluated. Various classification techniques used in the proposed work are as follows: (a) deep CNN, (b) bagging ensemble with SVM, (c) extreme learning machine (ELM), and (d) online sequential ELM (OS-ELM).

Deep CNN

CNN can successfully capture the spatial and temporal dependencies in an image through the application of relevant filters. The architecture performs a better fitting to the image dataset due to the reduction in the number of parameters involved and the re-usability of weights [43]. Since VGG is itself a CNN architecture, for the deep CNN model, a fully connected layer of size 1024 is added to the truncated VGG architecture followed by a softmax layer for classification. This gives us the most direct classification model where the feature extraction and classification are in the same CNN architecture. The deep CNN utilizes the fine-tuned weights and uses it to directly predict the output. Similar model has been used by authors in [11] where they have used fine-tuned Resnet-50 for chest CT scan classification and achieved an accuracy of 93%. Summary of various VGG16 truncation point accuracy evaluated on the validation set with SVM as classifier

Extreme learning machine

ELMs are single-hidden layer feedforward neural networks (SLFNs) that randomly choose hidden nodes and analytically determines the output weights of SLFNs through the generalized inverse operation of the hidden layer output matrices. The implementation of ELM is as described in [44]. The number of hidden nodes in the model is experimentally determined with the best-suited gamma (width multiplier for RBF distance). L2-normalized RBF activation function has been used. Experimentation has also been conducted with varying numbers of neurons in the hidden layer. Using validation set accuracy, it was observed that the highest accuracy was observed at 1000 hidden nodes. The performance of ELM is comparable to [45] where authors have used ELM to classify COVID-19 chest C-rays and have achieved an F1 score of 0.95.

Online sequential ELM

OS-ELM can learn data chunk by chunk with varying chunk size and provides faster sequential learning. The implementation of the model is the same as that described in [46]. It uses the idea of ELMs with a sequential determination of the output weights through the recursive least-squares (RLS) algorithm. OS-ELM consists of two phases, namely an initialization phase and a sequential learning phase. In the initialization phase, a base extreme learning machine model is trained using a small chunk of initial training data. For classification using OS-ELM, SLFN is implemented with a sigmoid activation function with 2500 hidden layers. As the model has very few hyperparameters, they have been optimized using a grid-search optimization.

Bagging ensemble with SVM

To improve the limited performance of the SVM (accuracy of 93.4%) due to the high complexity of time and space, the SVM ensemble with bagging is used. A single classifier may have a high test-error, but many small classifiers can produce a low test error and increase robustness because diversity compensates for error. For classification using the bagging SVM, the dataset is randomly divided into 10 parts. The individual classifiers are trained independently with the bootstrap technique and aggregated to make a joint decision by the deterministic averaging process. The proposed classifier model with “RBF” kernel and tuned hyperparameters is used as the base estimator. Bagging ensemble with SVM achieves the highest accuracy of 95.7% on the testing data. Due to the high accuracy of the bagging with the SVM model, it is the proposed classification method for COVID-19 screening. SVM also achieved exceptional results in biomedical studies such as [47, 48]. Performances analysis of feature selection techniques on validation set using SVM as classifier Performance parameters of different classifiers on testing data

Evaluation metrics

Confusion matrices for different classifiers are shown in Fig. 8. The classifiers are evaluated on the test set with 111 COVID-19 images and 97 non-COVID images. The features for the model are extracted using the truncated VGG16 model and selected using PCA. The screening performance of the model is assessed using generalized performance parameters derived from the confusion matrix. Table 6 put forth the generalized performance parameters, namely, true positive (TP), false positive (FP), true negative (TN), false negative (FN), AUC, accuracy (ACC), precision (PRE), sensitivity (S1), specificity (S2), negative predictive value (NPV), and F1 score (F1).
Fig. 8

Confusion matrices of the proposed methodology with different classifiers

Table 6

Performance parameters of different classifiers on testing data

ClassifierTPTNFPFNAUCPRENPVS1S2F1ACC
Bagging with SVM107925495.80.9550.9580.9630.9480.9590.957
ELM107889493.80.9220.9560.9630.9070.9420.937
OS-ELM107907494.90.9380.9570.9630.9270.9510.947
Deep CNN1038215889.50.8720.9110.9270.8450.8990.889

Experimentation

Testing environment

The proposed methodology is implemented on python software, run on a CPU. The system requirements are an Intel Core i7 processor with a 4 GB graphic card, a 64-bit operating system at 1.80 GHz, and 16 GB RAM.

Screening of COVID-19 based on different classifier

Figure 6 shows the convergence graph of training and validation accuracy of the transfer learning–based CNN model for the bagging ensemble classifier with SVM. Figure 5 shows the confusion matrices of the proposed architecture with and without fine-tuning of the VGG16 based feature extractor. The confusion matrices are obtained by evaluating the models on the test set with bagging SVM as the classifier.
Fig. 6

Convergence graph of accuracy vs epoch for proposed methodology (VGG16+PCA+bagging ensemble with SVM)

Convergence graph of accuracy vs epoch for proposed methodology (VGG16+PCA+bagging ensemble with SVM)

K-fold cross-validation

K-fold cross-validation divides the training set into k-subsets and trains k-models such that one of the subsets is left out while training the model. The accuracy of these k models is then averaged. The benefit of k-fold is that the entire dataset is used for both training and validation, and each sample is used for validation exactly once. The 10-fold cross-validation curve obtained for the bagging SVM is shown in Fig. 7. This is a 10-fold cross-validation curve plotted against the number of training examples. It compares the robustness of the model with experience and shows that the proposed model is well generalized. The average scores of 10-folds have been shown in the plot.
Fig. 7

Learning curve for proposed method using 10-fold cross-validation

Learning curve for proposed method using 10-fold cross-validation

Adversial defense

Deep learning models are often fooled with noise perturbations in the image. Such perturbations or attacks lead to miss classification of images. To defend the model against such noise attacks, a defense module has been designed. To remove noise from an image before prediction, three image denoiser have been applied namely total variation, Gaussian filter, and wavelet denoising. The prediction of all three denoised images is passed to an ensemble which finally classifies the image as shown in Fig. 8. On evaluating this module with the test set after adding random noise, the model gave an accuracy of 82.34%. Confusion matrices of the proposed methodology with different classifiers

Results

In the proposed work, the best performing model achieves an accuracy of 95.67% along with a precision of 96.83%. The area under the ROC curve (AUC) obtained is 95.8, as shown in Fig. 9. The proposed method aims to reduce the false-negative rate as much as possible since false-positive cases can potentially be identified in subsequent tests, but false-negative cases might not have that chance. The proposed model has a false negative rate of 4.33%, which is significantly lower than other COVID-19 CT scan screening models. The model proposed in this study achieves a very high accuracy of 95.67% on the testing data with a very low prediction time of 358 ms. This proves that deep learning-based approaches could be used to effectively and accurately screen COVID-19 at very large scales. Table 7 puts forth the comparative analysis of the proposed methodology with other existing techniques.
Fig. 9

ROC characteristics curve for the proposed methodology (VGG16+PCA+bagging ensemble with SVM)

Table 7

Comparative analysis of COVID-19 detection proposed methodology with techniques available in the literature on the used dataset

Sr. No.TechniquesDatasetPerformance evaluation
1.DECAPS + Peekaboo [27]Binary classification of total 746 chest CT images COVID-19 and non-COVID-19 [15]Accuracy - 87.6%, AUC- 0.961, and precision - 84.3%.
2.Resnet50 and VGG16 [13]Total 102 X-ray images of COVID-19-positive and -pneumonia patients [24], [49], [50], [14]Overall accuracy achieved is 89.2%.
3.AI methods (JCS and DenseNet169) [15]Binary classification of CT scan data into COVID (349 images) and non-COVID (463 CT images) [51]Accuracy - 0.83, AUC - 0.95, and F1 - 0.85.
4.Proposed methodology VGG16+PCA+Bagging Ensemble with SVMBinary classification (COVID-19 and non-COVID-19) using 702 CT scan images (344 COVID-19 images and 358 non-COVID images) [14], [24], and [15].Prediction time is 385ms, Accuracy - 95.7%, Precision - 95.8%, AUC - 0.958, and F1 score - 95.3%.
ROC characteristics curve for the proposed methodology (VGG16+PCA+bagging ensemble with SVM) Comparative analysis of COVID-19 detection proposed methodology with techniques available in the literature on the used dataset

Conclusion

A deep learning-based truncated VGG16 model is proposed in this study to screen COVID-19 patients using chest CT scans. The VGG16 architecture is fine-tuned and used to extract features from CT scan images. An interesting observation has been that pre-trained models are able to learn features very effectively with fine-tuning. The study demonstrates that VGG16 outperforms other models on biomedical image feature extraction. Another important observation has been that truncation improves the model’s performance on the limited dataset. The performance also improves on applying dimensionality reduction techniques, indicating a high correlation among features that need to be removed to boost the performance of the classifier. It was experimentally found that PCA performs much better than auto-encoders for biomedical image features, as has been verified by several other studies. Finally, a comparison of various popular classification techniques shows us that a bagging ensemble of SVM gives the best results and outperforms direct CNN classification, ELM, and OS-ELM. Our study addresses the challenges associated with the limited and poor quality of COVID-19 radiology data. The study has proposed techniques like transfer learning, fine-tuning, model-truncation, image augmentation, and feature-reduction to overcome these. This should be helpful for practitioners aiming to use these datasets for their research and development. Furthermore, the importance of various pre-processing techniques has also been highlighted. While comparing different pre-trained models, it was found that both VGG16 and VGG19 gave great results within the constraints of the small dataset. While deeper networks were found to struggle, they will perform better when larger datasets are available. For currently available datasets, we conclude that VGG16 with appropriate truncation and fine-tuning gives the best feature maps. These features give good results when classified directly but the results are found to improve significantly when a feature selector like PCA is applied before classification. With the quality of available COVID-19 data increasing, clinically verified, and trusted deep learning models may be developed for fast diagnosis of COVID-19. The superior performance of the deep models may result in AI-based diagnosis for various diseases especially in times of outbreak where rapid screening and early identification are crucial for effective containment. As the quality of available data increases, deep models can be very effective for diagnosis. Future works on the study would focus on making the model interpretable and subjective in classification with a severity score which may be useful in the case of screening. Furthermore, a region of infection marking may assist the medical staff to treat the infection.
  21 in total

1.  A fast and accurate online sequential learning algorithm for feedforward networks.

Authors:  Nan-Ying Liang; Guang-Bin Huang; P Saratchandran; N Sundararajan
Journal:  IEEE Trans Neural Netw       Date:  2006-11

2.  COVID-19 Detection Through Transfer Learning Using Multimodal Imaging Data.

Authors:  Michael J Horry; Subrata Chakraborty; Manoranjan Paul; Anwaar Ulhaq; Biswajeet Pradhan; Manas Saha; Nagesh Shukla
Journal:  IEEE Access       Date:  2020-08-14       Impact factor: 3.367

3.  Automated Deep Transfer Learning-Based Approach for Detection of COVID-19 Infection in Chest X-rays.

Authors:  N Narayan Das; N Kumar; M Kaur; V Kumar; D Singh
Journal:  Ing Rech Biomed       Date:  2020-07-03

4.  Text Data Augmentation for Deep Learning.

Authors:  Connor Shorten; Taghi M Khoshgoftaar; Borko Furht
Journal:  J Big Data       Date:  2021-07-19

5.  Truncated inception net: COVID-19 outbreak screening using chest X-rays.

Authors:  Dipayan Das; K C Santosh; Umapada Pal
Journal:  Phys Eng Sci Med       Date:  2020-06-25

6.  Systematic review of artificial intelligence techniques in the detection and classification of COVID-19 medical images in terms of evaluation and benchmarking: Taxonomy analysis, challenges, future solutions and methodological aspects.

Authors:  O S Albahri; A A Zaidan; A S Albahri; B B Zaidan; Karrar Hameed Abdulkareem; Z T Al-Qaysi; A H Alamoodi; A M Aleesa; M A Chyad; R M Alesa; L C Kem; Muhammad Modi Lakulu; A B Ibrahim; Nazre Abdul Rashid
Journal:  J Infect Public Health       Date:  2020-07-01       Impact factor: 3.718

7.  Real-time RT-PCR in COVID-19 detection: issues affecting the results.

Authors:  Alireza Tahamtan; Abdollah Ardebili
Journal:  Expert Rev Mol Diagn       Date:  2020-04-22       Impact factor: 5.225

8.  Classification of COVID-19 patients from chest CT images using multi-objective differential evolution-based convolutional neural networks.

Authors:  Dilbag Singh; Vijay Kumar; Manjit Kaur
Journal:  Eur J Clin Microbiol Infect Dis       Date:  2020-04-27       Impact factor: 3.267

9.  COVID19XrayNet: A Two-Step Transfer Learning Model for the COVID-19 Detecting Problem Based on a Limited Number of Chest X-Ray Images.

Authors:  Ruochi Zhang; Zhehao Guo; Yue Sun; Qi Lu; Zijian Xu; Zhaomin Yao; Meiyu Duan; Shuai Liu; Yanjiao Ren; Lan Huang; Fengfeng Zhou
Journal:  Interdiscip Sci       Date:  2020-09-21       Impact factor: 2.233

Review 10.  Chest CT manifestations of new coronavirus disease 2019 (COVID-19): a pictorial review.

Authors:  Zheng Ye; Yun Zhang; Yi Wang; Zixiang Huang; Bin Song
Journal:  Eur Radiol       Date:  2020-03-19       Impact factor: 7.034

View more
  12 in total

1.  A Review on Deep Learning Techniques for the Diagnosis of Novel Coronavirus (COVID-19).

Authors:  Md Milon Islam; Fakhri Karray; Reda Alhajj; Jia Zeng
Journal:  IEEE Access       Date:  2021-02-10       Impact factor: 3.367

2.  Non-iterative learning machine for identifying CoViD19 using chest X-ray images.

Authors:  Sahil Dalal; Virendra P Vishwakarma; Varsha Sisaudia; Parul Narwal
Journal:  Sci Rep       Date:  2022-07-13       Impact factor: 4.996

3.  Machine Learning-Based Research for COVID-19 Detection, Diagnosis, and Prediction: A Survey.

Authors:  Yassine Meraihi; Asma Benmessaoud Gabis; Seyedali Mirjalili; Amar Ramdane-Cherif; Fawaz E Alsaadi
Journal:  SN Comput Sci       Date:  2022-05-12

4.  An integrated feature frame work for automated segmentation of COVID-19 infection from lung CT images.

Authors:  Deepika Selvaraj; Arunachalam Venkatesan; Vijayalakshmi G V Mahesh; Alex Noel Joseph Raj
Journal:  Int J Imaging Syst Technol       Date:  2020-11-23       Impact factor: 2.177

Review 5.  Medical image processing and COVID-19: A literature review and bibliometric analysis.

Authors:  Rabab Ali Abumalloh; Mehrbakhsh Nilashi; Muhammed Yousoof Ismail; Ashwaq Alhargan; Abdullah Alghamdi; Ahmed Omar Alzahrani; Linah Saraireh; Reem Osman; Shahla Asadi
Journal:  J Infect Public Health       Date:  2021-11-17       Impact factor: 3.718

Review 6.  Supervised and weakly supervised deep learning models for COVID-19 CT diagnosis: A systematic review.

Authors:  Haseeb Hassan; Zhaoyu Ren; Chengmin Zhou; Muazzam A Khan; Yi Pan; Jian Zhao; Bingding Huang
Journal:  Comput Methods Programs Biomed       Date:  2022-03-05       Impact factor: 7.027

7.  A novel explainable COVID-19 diagnosis method by integration of feature selection with random forest.

Authors:  Mehrdad Rostami; Mourad Oussalah
Journal:  Inform Med Unlocked       Date:  2022-04-06

8.  Statistical analysis of COVID-19 infection severity in lung lobes from chest CT.

Authors:  Mehdi Yousefzadeh; Mozhdeh Zolghadri; Masoud Hasanpour; Fatemeh Salimi; Ramezan Jafari; Mehran Vaziri Bozorg; Sara Haseli; Abolfazl Mahmoudi Aqeel Abadi; Shahrokh Naseri; Mohammadreza Ay; Mohammad-Reza Nazem-Zadeh
Journal:  Inform Med Unlocked       Date:  2022-04-01

9.  Deep Learning-Based Approaches to Improve Classification Parameters for Diagnosing COVID-19 from CT Images.

Authors:  Huseyin Yasar; Murat Ceylan
Journal:  Cognit Comput       Date:  2021-07-15       Impact factor: 4.890

10.  DenseNet Convolutional Neural Networks Application for Predicting COVID-19 Using CT Image.

Authors:  Najmul Hasan; Yukun Bao; Ashadullah Shawon; Yanmei Huang
Journal:  SN Comput Sci       Date:  2021-07-23
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.