Literature DB >> 35821878

Detection of COVID-19 using deep learning techniques and classification methods.

Çinare Oğuz1, Mete Yağanoğlu1.   

Abstract

Since the patient is not quarantined during the conclusion of the Polymerase Chain Reaction (PCR) test used in the diagnosis of COVID-19, the disease continues to spread. In this study, it was aimed to reduce the duration and amount of transmission of the disease by shortening the diagnosis time of COVID-19 patients with the use of Computed Tomography (CT). In addition, it is aimed to provide a decision support system to radiologists in the diagnosis of COVID-19. In this study, deep features were extracted with deep learning models such as ResNet-50, ResNet-101, AlexNet, Vgg-16, Vgg-19, GoogLeNet, SqueezeNet, Xception on 1345 CT images obtained from the radiography database of Siirt Education and Research Hospital. These deep features are given to classification methods such as Support Vector Machine (SVM), k Nearest Neighbor (kNN), Random Forest (RF), Decision Trees (DT), Naive Bayes (NB), and their performance is evaluated with test images. Accuracy value, F1-score and ROC curve were considered as success criteria. According to the data obtained as a result of the application, the best performance was obtained with ResNet-50 and SVM method. The accuracy was 96.296%, the F1-score was 95.868%, and the AUC value was 0.9821. The deep learning model and classification method examined in this study and found to be high performance can be used as an auxiliary decision support system by preventing unnecessary tests for COVID-19 disease.
© 2022 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  COVID-19; Classification; Deep learning; ResNet

Year:  2022        PMID: 35821878      PMCID: PMC9263717          DOI: 10.1016/j.ipm.2022.103025

Source DB:  PubMed          Journal:  Inf Process Manag        ISSN: 0306-4573            Impact factor:   7.466


Introduction

The COVID-19 pandemic has halted the economy around the world and has caused congestion in healthcare institutions, especially intensive care units (Firouzi et al., 2021, Fleuren, Dam et al., 2021, Garcia et al., 2021). It requires staying in intensive care in the later stages of the COVID-19 disease (Fleuren et al., 2021). It is tried to reach the diagnosis of COVID-19 by performing Polymerase Chain Reaction (PCR) test and rapid antigen test, by taking computed tomography (CT) image or X-ray image. CT images are created by combining X-ray images taken from multiple angles of the determined area of the patient’s body. The PCR test result used to diagnose COVID-19 may be negative, especially in the early stage of the disease or when the viral load is low. Taking CT images from the patient at this stage plays an important role in the diagnosis of COVID-19 (Kang, Li, & Zhou, 2020). Fang et al. (2020) indicated may lead to negative PCR test result due to the immature development of nucleic acid detection technology, differences in detection rate from different manufacturers and inappropriate clinical sampling. Fang et al. (2020) found the sensitivity of the PCR test to be 71% in the early detection of COVID-19, and 98% of the sensitivity of the detection of COVID-19 by taking CT images. CT imaging is more preferred than X-ray imaging in the diagnosis of COVID-19. Because in the early stages of COVID-19 disease, the ground glass image in the lung may not be seen in X-ray images. The sensitivity of X-ray imaging is lower than that of CT imaging (Islam, Karray, Alhajj, & Zeng, 2021). This pandemic poses a challenge for healthcare workers around the world. It takes time to evaluate the test results of many patients. The need for solutions that can provide clinical support for the treatment of COVID-19 patients has increased (Romeo & Frontoni, 2022). Image analysis on medical images with decision support systems can provide accurate and rapid diagnosis of disease to cope with a large number of patient demands. For example, it may take up to 10 min for radiologists to manually evaluate CT images of a patient, while image analysis with decision support systems requires only a few seconds. COVID-19 was spreading very fast (Fleuren, Tonutti et al., 2021); Although it causes serious organ dysfunction such as pneumonia, kidney failure, sometimes it results in death. Therefore, early diagnosis of COVID-19 cases is necessary. Since the patient is not quarantined during the conclusion of the PCR test used in the diagnosis of COVID-19, the disease continues to spread.

Research objective

In this research, we created hybrid models by integrating deep learning models, which is a popular branch of machine learning, and traditional machine learning methods. Among the models created, a new high-performance COVID-19 detection system was developed with the hybrid model of ResNet-50 and SVM. In our proposed hybrid model framework, there are 2 modules including feature extraction with deep learning method and classification with traditional machine learning method. We evaluated the proposed hybrid model on a new dataset of lung CT images obtained from a hospital setting and labeled by a radiologist. As shown in the experimental results, the proposed hybrid model (ResNet-50SVM) achieved higher accuracy than the classical ResNet-50 method. The main purpose of this article is to reduce the duration and amount of transmission of the disease by shortening the diagnosis time by taking CT images from patients who were admitted to the hospital due to the suspicion of COVID-19. In addition, it is aimed to provide a decision support system to radiologists in the detection of COVID-19.

Related works

In this section literature summary is discussed based on deep learning-based support systems for COVID-19 detection. The works in the literature were analyzed one by one based on the research methodologies and experimental results. These works are grouped according to the types of medical images used, such as CT and X-ray.

COVID-19 detection using X-ray images

A summary of studies using X-ray images in the literature is given in Table 1. Ucar and Korkmaz (2020) proposed a Bayes-SqueezeNet based COVIDiagnosis-Net for the diagnosis of COVID-19 on X-ray images. Dataset consisted of 45 COVID-19 cases, 1203 normal and 1591 non-COVID-19 pneumonia cases images. Dataset consists of a combination of three open-source datasets. According to the results obtained, the proposed model has an accuracy rate of 98.26% and an F1-score value of 98.25%.
Table 1

Summary table: COVID-19 detection studies using X-ray images.

StudyMethodNumber of imagesData sourcePerformance
Ucar and Korkmaz (2020)Bayes-SqueezeNet1203 Normal 1591 Pneumonia 45 COVID-19Open source98.26 Accuracy

Panwar, Gupta, Siddiqui, Morales-Menendez, and Singh (2020)nCOVNet142 Normal 142 COVID-19Open source88% Accuracy

Oh, Park, and Ye (2020)Patch-based CNN8851 Normal 6012 Pneumonia 180 COVID-19Open source88.9% Accuracy

Mahmud, Rahman, and Fattah (2020)CovXNet1583 Normal 2780 Bacterial pneumonia 1493 Viral pneumonia 305 COVID-19Open source90.2% Accuracy

Panahi, Rafiei, and Rezaee (2021)FCOD505 Normal 435 COVID-19Open source96% Accuracy
Panwar et al. (2020) presented a deep learning neural network-based nCOVnet to detect COVID-19 patients on X-ray images. In the dataset, 142 X-ray images were selected as COVID-19 positive and 142 X-ray images obtained from the Kaggle dataset were selected as COVID-19 negative. 30% of this data set was used as test data. The proposed model detected COVID-19 positive images with an accuracy rate of 97%, while the overall accuracy was recorded as 88%. Summary table: COVID-19 detection studies using X-ray images. Oh et al. (2020) proposed a patch-based CNN with few trainable parameters. The dataset includes X-ray images of 180 COVID-19, 6012 pneumonia and 8851 normal cases. The proposed model has an accuracy rate of 88.9% in the experiment. Mahmud et al. (2020) used X-ray images as a data set for the detection of COVID-19 and other pneumonia cases. A deep learning model named CovXNet has been proposed. The data set consists of 305 COVID-19, 2780 bacterial pneumonia, 1493 viral pneumonia and 1583 normal images. An accuracy rate of 97.4% was determined for the detection of COVID-19 images, and an accuracy rate of 90.2% for the general classification. Panahi et al. (2021) proposed an inception architecture based Fast COVID-19 Detector for rapid detection of COVID-19 on X-ray images. The dataset consists of 940 open source X-ray images containing COVID-19 and normal cases. 80% of this data set is reserved for training and 20% for testing. The accuracy rate in detecting COVID-19 was 96%.

COVID-19 detection using CT images

A summary of studies using CT images in the literature is given in Table 2. Singh, Kumar, Kaur, et al. (2020) proposed a CNN for detect COVID-19 patients. The dataset consists of 75 COVID-19 positive and 75 COVID-19 negative CT images. Experiments were conducted using different variations while separating the data set. The best variation in the proposed model was found to be 90% for training and 10% for testing. The proposed model achieved 90.72% specificity, 90.7% sensitivity, 93.25% accuracy and 89.96% F1 score.
Table 2

Summary table: COVID-19 detection studies using CT images.

StudyMethodNumber of imagesData sourcePerformance
Wu et al. (2020)COVNet1325 Normal 1735 Pneumonia 1296 COVID-19Multiple hospital environment90% sensitivity

Zheng et al. (2020)DeCOVNet630Multiple hospital environment90.7% sensitivity

Singh et al. (2020)MODE-CNN75 Normal 75 COVID-19Open source database90.7% sensitivity

Pham (2020)DenseNet-201397 Normal 349 COVID-19Open source database96.2% accuracy

Wu, Chen, Zhong, Wang, and Shi (2021)COVIDAL342 Normal 316 Pneumonia 304 COVID-19Open source database86.6% accuracy

Rohila, Gupta, Kaul, and Sharma (2021)ReCOV-111110MosMedData94.9% accuracy

Wang et al. (2021)Modified Inception V31065Multiple hospital environment85.2% accuracy

Li, Yang, Liang, and Wu (2021)CheXNet397 Normal 349 COVID-19Open source database87% accuracy

Shui-Hua et al. (2022)DRAPNet306 Normal 281 Pneumonia 284 COVID-19 293 SPTMultiple hospital environment95.49% F1-Score

Wang, Zhang and Zhang (2021)DSSAE306 Normal 281 Pneumonia 284 COVID-19 293 SPTMultiple hospital environment92.32% F1-Score
Wu et al. (2020) offered a model called COVNet based on the ResNet50 deep learning technique to reach the diagnosis of COVID-19. The data set consisting of CT images includes 4536 images, 1296 of which are COVID-19, 1735 are pneumonia and 1325 are normal. 90% of the dataset is reserved for training and 10% for testing. According to the experimental results, the proposed model achieved 90% sensitivity, 96% specificity and 96% AUC. Summary table: COVID-19 detection studies using CT images. Zheng et al. (2020) proposed a 3D deep CNN type model (DeCoVNet) for the detection of COVID-19. The dataset consists of 630 CT images collected from the hospital environment. 80% of the data set is reserved for training and 20% for testing. In the experiments, 90.1% accuracy, 90.7% sensitivity, 91.1% specificity, 95.9% AUC value were obtained. Wu et al. (2021) offered deep learning based COVID-AL. The dataset consists of CT images collected from the China Thorax CT Image Review Consortium (CC-CCII). In this data set, 304 COVID-19, 316 pneumonia and 342 normal case images were used. An accuracy rate of 86.6% was founded. Pham (2020) proposed DenseNet-201 transfer learning model for COVID-19 detection. The performance of the model was evaluated on 746 CT images obtained from the open source dataset. The dataset contains 349 COVID-19 images and 397 normal images. Data augmentation was not applied. An accuracy value of 96.2% was obtained in the experimental results. Wang, Kang et al. (2021) presented by changing the architecture of InceptionV3, which is the transfer-learning model for COVID-19 detection. The performance of the model was evaluated on 1065 CT images. In the experimental results, an accuracy value of 85.2% was obtained. Li et al. (2021) proposed a pre-trained CheXNet model for COVID-19 detection. The performance of the model was evaluated on 746 CT images obtained from the open source dataset. The dataset contains 349 COVID-19 images and 397 normal images. In the experimental results, an accuracy value of 87% was obtained. Shui-Hua et al. (2022) proposed a new deep rank-based mean pooling network (DRAPNet) model, based on the rank-based mean pooling module (NRAPM) and inspired by the VGG network, using CT images for COVID-19 detection. The performance of the model was evaluated on a multi-hospital dataset containing 1164 CT images. The dataset contains 284 COVID-19, 281 pneumonia, 293 pulmonary tuberculosis and 306 normal images. According to the test results, the proposed model reached an F1-score value of 95.49. Rohila et al. (2021) proposed a model called ReCOV-101 for detection of COVID-19. The model is a DCNN model using ResNet-101. Preprocessing such as segmentation has been applied on full chest CT scans. In the experimental results, the accuracy value of 94.9% was reached. Wang, Zhang et al. (2021) proposed a deep-stack sparse auto encoder (DSSAE) model for COVID-19 detection. Features are obtained using two-dimensional fractional Fourier entropy. The performance of the model was evaluated on a multi-hospital dataset containing 1164 CT images. The dataset includes 284 COVID-19, 281 pneumonia, 293 s pulmonary tuberculosis(SPT) and 306 normal images. According to the experimental results, the proposed model achieved an F1-score value of 92.32. In addition, using fusion of CT and X-ray images were obtained by Zhang, Zhang, Zhang, and Wang (2021). They proposed a multi-input deep convolutional attention network (MIDCAN) model for COVID-19 detection. The performance of the model was evaluated on a multi-hospital dataset containing mixed CT and X-ray images. The dataset contains 42 COVID-19 and 44 normal images. According to the experimental results, the proposed model achieved an accuracy value of 98.02%. In the detection of COVID-19 disease, experts prefer CT imaging more than X-ray imaging. Because in the early stages of COVID-19 disease, the ground glass image in the lung may not be seen in X-ray images. The sensitivity of X-ray imaging is lower than that of CT imaging (Islam et al., 2021). Therefore, CT images were used in this study. In the studies in the literature, deep learning models have been used alone. In this study, performance was increased by using deep learning models and classification methods together. Also the run times of the methods used in most of the studies in the literature are not mentioned.

Material and method

The block diagram of this study is shown in Fig. 1. Patients who come to Siirt Research and Training Hospital with the suspicion of COVID-19 between May 2020 and June 2020 are accepted as participants. Deep features are extracted from different deep learning models (ResNet-50, ResNet-101, AlexNet, Vgg-16, Vgg-19, GoogLeNet, SqueezeNet and Xception) by using the images reserved for training from the dataset. By using these deep features in various classification methods (SVM, kNN, RF, DT and NB), it is tried to find the deep learning model and classification method pair with the highest performance. In this study, the fully connected layer of all deep learning models is replaced with a new fully connected layer whose output number is equal to the number of classes in the new dataset. In the classification process, it was done with various traditional machine learning methods instead of softmax.
Fig. 1

Block diagram.

In the input layer of the deep learning model, images are adapted to the model. In the convolution layer, feature maps are created with the convolution filters of the model. In the first convolution layer, the edges of the input are calculated. The edges show the high-frequency regions of the input. After these calculations, the padding method is applied because the output size will be different from the input size. Thus, the output and input sizes are synchronized. Deep features are extracted from the fc layer of the model. Learning takes place by using the deep features and class labels of the training data in different classification methods. Then, the deep features of the images left for testing from the dataset are extracted with deep learning models. The estimation process is performed by using the extracted deep features in the trained classification methods. Finally, by comparing the class labels of the test data with the predicted class labels, deep learning models and classification methods are evaluated with performance metrics such as accuracy, sensitivity, specificity, F1 score, runtime and AUC. Block diagram.

Data set

Ataturk University Non-Invasive Clinical Research Ethics Committee approval was obtained for the data set used in this study. With this approval, COVID-19 positive and COVID-19 negative CT data from patients who applied to Siirt Education and Research Hospital COVID-19 Diseases polyclinics between May 2020 and June 2020 were obtained through the automation system and a data set was created. CT images were taken from adults over 18 years of age. Data set was obtained from 1275 patients in total. 53% of patients who are positive for COVID-19 consist of male patients. One hundred of these patients are intensive care patients, and CT images were taken again from the same patient on one day, seven days and 14 days, depending on the length of stay for each patient. 72% of intensive care patients consist of patients over 54 years of age. A data set was created using a total of 1345 CT images. While 738 of them were negative for COVID-19, 607 of them were positive for COVID-19. In the study, 270 (20%) of the CT images in the data set were used for testing. Details of the dataset are given in Table 3.
Table 3

Details of dataset.

COVID-19 positiveCOVID-19 negative
Training set485590
Test set122148
Total607738
Details of dataset.

Deep learning models

AlexNet

Maximum pooling(max pool) is performed on 3 × 3 pixel windows. For maximum pooling, the stride is set to [2 2], padding is set to [0 0 0 0]. The activation function (RELU) is applied after each convolutional layer (conv layer) and to the output of the fully connected layer (fc layer). The first conv layer contains convolution filters with 96 cores of 11 × 11 size. In the first conv layer, the stride for the filter is set to [4 4] and padding is set to [0 0 0 0]. Normalization and pooling operations are applied to the outputs of the first and second conv layers. No pooling or normalization is applied between the third, fourth, and fifth conv layers. The second conv layer contains convolution filters with 256 cores of 5 × 5 size. In the second conv layer, the stride for the filter is set to [1 1] and padding is set to [2 2 2 2]. The third conv layer contains convolution filters with 384 cores of 3 × 3 size. In the third conv layer, stride for the filter is set to [1 1] and padding is set to [1 1 1 1]. The fourth conv layer has convolution filters with 384 cores of 3 × 3 size. In the fourth conv layer, the stride for the filter is set to [1 1] and padding is set to [1 1 1 1]. The fifth conv layer has convolution filters with 256 cores of 3 × 3 size. In the fifth conv layer, the stride for the filter is set to [1 1] and padding is set to [1 1 1 1]. The first two fc layers each contain 4096 channels and the last fc layer contains 1000 channels.

ResNet

The image input size of the model is 224 × 224. Maximum pooling is performed on 3 × 3 pixel windows. For maximum pooling, the stride is set to [2 2], padding is set to [1 1 1 1]. The 7 × 7 average pooling is done after the fifth conv layer. RELU is implemented after each conv layer. The first conv layer contains convolution filters with 64 cores of 7 × 7 size. In the first conv layer, the stride for filters is set to [2 2] and padding is set to [3 3 3 3]. The second conv layer takes the output of RELU and pooling operations after the first conv layer as input. It includes convolution filters with 64 cores of 1 × 1 and 3 × 3 sizes and 256 cores of 1 × 1 size. In the second conv layer, the stride for filters is set to [1 1], padding is set to [0 0 0 0] for 1 × 1 filters and ‘same’ for 3 × 3 filters. The third conv layer takes the output of RELU and normalization operations after the second conv layer as input. It includes convolution filters with 128 cores of 1 × 1 and 3 × 3 sizes and 512 cores of 1 × 1 size. In the third conv layer, the stride for filters is set to [1 1], padding is set to [0 0 0 0] for 1 × 1 filters and ‘same’ for 3 × 3 filters. The fourth conv layer takes the output of the RELU and normalization operations after the third conv layer as input. It includes convolution filters with 256 cores of 1 × 1 and 3 × 3 sizes and 1024 cores of 1 × 1 size. In the fourth conv layer, the stride for filters is set to [1 1], padding is set to [0 0 0 0] for 1 × 1 filters and ‘same’ for 3 × 3 filters. The fifth conv layer takes the output of the RELU and normalization operations after the fourth conv layer as input. It includes convolution filters with 512 cores of 1 × 1 and 3 × 3 sizes and 2048 cores of 1 × 1 size. In the fifth conv layer, the stride for filters is set to [1 1], padding is set to [0 0 0 0] for 1 × 1 filters and ‘same’ for 3 × 3 filters.

Vgg

Vgg takes 224 × 224 image input. There are five max pool filters placed between conv layers for size reduction. Maximum pooling is performed on 2 × 2 pixel windows. For max pool, the stride is set to [2 2], padding is set to [0 0 0 0]. In the dropout layer, the outputs are reduced by 50% (drop value 0.5). The last layer is the softmax layer. The stride for filters in conv layers is set to [1 1] and padding is set to [1 1 1 1]. The first conv layer contains convolutional filters with 64 cores of 3 × 3 size. The second conv layer takes the output of RELU and pooling operations after the first conv layer as input. Includes convolution filters with 128 core of 3 × 3 size. The third conv layer takes the output of RELU and pooling operations after the second conv layer as input. It includes convolution filters with 256 cores of 3 × 3 size. The fourth conv layer takes the output of RELU and pooling operations after the third conv layer as input. It includes convolution filters with 512 core of 3 × 3 size. The fifth conv layer takes the output of RELU and pooling operations after the fourth conv layer as input. Includes convolution filters with 512 core of 3 × 3 size. After the fifth conv layer, the final pooling takes place. Its output is given to the fc layer. RELU and dropout operations are applied between fc layers. The second conv layer takes the output of RELU and pooling operations after the first conv layer as input. This layer contains convolution filters with 128 cores of 3 × 3 size. The third conv layer takes the output of RELU and pooling operations after the second conv layer as input. This layer contains convolution filters with 256 cores of 3 × 3 size. The fourth conv layer takes the output of RELU and pooling operations after the third conv layer as input. This layer contains convolutional filters with 512 cores of 3 × 3 size. The fifth conv layer takes the output of RELU and pooling operations after the fourth conv layer as input. This layer contains convolutional filters with 512 cores of 3 × 3 size. After the fifth conv layer, the final pooling takes place. Its output is given to the fc layer. RELU and dropout operations are applied between fc layers.

GoogLeNet

The most obvious difference from the GoogLeNet deep learning model from other models is the Inception module. The basic logic of the Inception module is to reasonably reduce the size where the computational requirements would increase too much. Size reduction is performed with filter elements of different sizes such as 1 × 1, 3 × 3, 5 × 5 in this module. GoogLeNet’s image input size is 224 224. An alternative parallel max pool path is added at each stage. A total of five pooling processes are applied. Maximum pooling is performed on 3 × 3 pixel windows. For max pool operations, the stride is set to [2 2], padding is set to [0 1 0 1]. The first conv layer contains convolution filters with 64 cores of 7 × 7 size. In the first conv layer, the stride is set to [2 2] for the filters and the padding is set to [3 3 3 3]. The second conv layer takes the output of RELU, maximum pooling and normalization operations after the first conv layer as input. It includes convolution filters with 64 cores of size 1 × 1 and 192 cores of size 3 × 3. In the second conv layer, the stride is set to [1 1] for 1 × 1 filters, padding is set to [0 0 0 0] and the stride is set to [1 1] for 3 × 3 filters, padding is set to [1 1 1 1]. After the second conv layer, RELU takes the output of max pool and normalization operations as input to the inception module. After the second conv layer comes 9 inception modules. In Inception modules, the stride is set to [1 1], padding is set to [0 0 0 0] for 1 × 1 filters, [1 1 1 1] for 3 × 3 filters, and [2 2 2 2] for 5 × 5 filters. Maximum pooling is applied to the output of the second inception module and the output of the seventh inception module. The fc layer takes the output of 7 × 7 average pooling and dropout (0.4 drop value) operations after the last inception module as input.

SqueezeNet

Its different side from other deep learning models is fire modules. This module consists of an expansion layer consisting of a mixture of 1 × 1 and 3 × 3 convolution filters and a compression conv layer with only 1 × 1 convolution filters (Iandola et al., 2016). The image input size of the network is 227 × 227 (Bhole & Kumar, 2020). The SqueezeNet architecture starts with the independent conv layer. Then, the maximum pooling process takes place in the size of 3 × 3. For maximum pooling, the stride is set to [2 2], padding is set to [1 1 1 1]. Next comes 8 fire modules and conv layers. Two more max pool operations occur between fire modules. The stride is set to [2 2] for max pool between fire modules and padding is set to [0 1 0 1]. Finally, it ends with the average pooling process. In the dropout layer, the outputs are reduced by 50% (0.5 drop value). The first conv layer contains convolution filters with 64 cores of 3 × 3 size. For filters in the first conv layer, the stride [2 2] is set to padding is set to [0 0 0 0]. The stride is set to [1 1] for filters in fire modules, padding is set to [0 0 0 0] for filters of 1 × 1 size and [1 1 1 1] for filters of 3 × 3 size. The second conv layer takes the output of the dropout process after 8 fire modules as input. It contains convolution filters with 1000 cores of 1 × 1 size. In the second conv layer, the stride is set to [1 1] for the filters, padding is set to [0 0 0 0]. The average pooling layer takes the output of the RELU operation after the second conv layer as input.

Xception

The image input size of the network is 229 229 (Chollet, 2017). The model includes 14 block conv layers. Maximum pooling is performed on 3 × 3 pixel windows. For maximum pooling operations, stride is set to [2 2], padding is set to ‘same’. The first block conv layer contains convolution filters with 32 cores of 3 × 3 size and 64 cores of 3 × 3 size. In the first block conv layer, the stride is set to [2 2] for 32-core filters, the stride is set to [1 1] for 64-core filters, and padding is set to [0 0 0 0]. The second block conv layer takes the output of the RELU and normalization operations after the first conv layer as input. It includes convolution filters with 64-core of 3 × 3 size and 128-core of 1 × 1, 3 × 3 sizes. In the second block conv layer, stride is set to [1 1] for 3 × 3 filters, padding is set as ‘same’, stride is set as [1 1] for 1 × 1 filters, padding is set as [0 0 0 0]. The third block conv layer takes the output of RELU, normalization and pooling operations after the second block conv layer as input. It includes convolution filters with 128 cores of 3 × 3 size and 256 cores of 1 × 1, 3 × 3 sizes. In the third block conv layer, the stride is set to [1 1] for 3 × 3 filters, padding is set as ‘same’, stride is set to [1 1] for filters of 1 × 1 size, padding is set as [0 0 0 0]. The fourth block conv layer takes the output of RELU, normalization and pooling operations after the third block conv layer as input. It includes convolution filters with 256 cores of 3 × 3 size and 728 cores of 1 × 1, 3 × 3 sizes. In the fourth block conv layer, stride is set to [1 1] for 3 × 3 sized filters, padding is set as ‘same’, stride is set to [1 1] for 1 × 1 sized filters, padding is set to [0 0 0 0]. The fifth block conv layer takes the output of RELU, normalization and pooling operations after the fourth block conv layer as input. It includes convolution filters with 728 cores of 1 × 1 and 3 × 3 sizes. In the fifth block conv layer, stride is set to [1 1] for 3 × 3 size filters, padding is set as ‘same’, for 1 × 1 size filters stride is set as [1 1], padding is set as [0 0 0 0]. The sixth block conv layer takes the output of RELU and normalization operations after the fifth block conv layer as input. It includes convolution filters with 728 cores of 1 × 1 and 3 × 3 sizes. In the sixth block conv layer, the stride is set as [1 1] for 3 × 3 filters, padding is set as ‘same’, stride is set as [1 1] for filters of 1 × 1 size, padding is set as [0 0 0 0]. Until the thirteenth block conv layer, the other block conv layers perform the same operations and contain convolution filters of the same size. The thirteenth block conv layer takes as input the RELU and normalization operations after the twelfth block conv layer. It includes convolution filters with 728 cores of 3 × 3 size and 1024 cores of 1 × 1, 3 × 3 sizes. In the thirteenth block conv layer, the stride is set to [1 1] for 3 × 3 filters, padding is set as ‘same’, for 1 × 1 size filters, stride is set to [1 1], padding is set as [0 0 0 0]. The fourteenth block conv layer takes the output of the RELU, normalization and pooling operations after the thirteenth block conv layer as input. It includes 1536 cores of 1 × 1, 3 × 3 sizes and convolution filters with 2048 cores of 1 × 1 size. In the fourteenth block conv layer, stride is set to [1 1] for 3 × 3 filters, padding is set as ‘same’, stride is set as [1 1] for 1 × 1 size filters, padding is set as [0 0 0 0].

Classification methods

k nearest neighbor classification method

According to the k Nearest Neighbor (kNN) algorithm, the number of neighbor’s k is determined. The number of neighbors is not fixed. The choice of the best k varies according to the data. In binary classifications, it is better to choose k as an odd number to avoid binding decisions. In the data set in our study, the k value was tested for all values from 1 to 10 and the best result was determined as “k 5”. The distance function is calculated between the samples in the training set and the test data. The Euclidean method is used to calculate the distance function (Hu, Huang, Ke, & Tsai, 2016). The distance function is given in Eq. (1). Then the obtained distances are sorted in ascending order and the first k samples are selected. The most common class among these examples is considered to be the class of test data.

Support vector machines

Support Vector Machines (SVM) are the separation and classification of points on the plane by a line or hyperplane. While creating the SVM decision surface, it tries to maximize the distance between the two classes. Between these planes, there is only one hyperplane with a maximum boundary. The points that constrain the boundary width are called support vectors. The support vector algorithm tries to minimize the training error by classifying with the separating hyperplane with the largest boundary width. It aims to find a linear discriminating function with the largest margin that separates the classes from each other. For non-linearly separable samples, the samples can be separated linearly into another higher-dimensional. A decision boundary is created with the training data set and the test data is classified according to this boundary (Zhou, 2021).

Decision trees

The representation of the information presented in the Decision Trees (DT) can be easily understood. DT not only shows decisions but also contains explanations of decisions. DT starts at the root of the tree to guess the class label of an instance. Root Node: It represents the entire population. They were further divided into two or more homogeneous clusters. Initially, the entire training set is considered the root. Root property values are compared with the properties of the sample. That is, to follow the branch corresponding to this value and jump to the next node. Samples are distributed recursively based on property values. While the features are placed in the tree, determining them as root or internal nodes is done using some statistical approaches. These approaches are; entropy is a measure of randomness (Charbuty & Abdulazeez, 2021). When the probability is 0 or 1, the result is certain. Since there is no randomness, the entropy value is zero. The entropy calculation for a single feature is calculated as in Eq. (2). S is the current sample, c is the number of classes, and pi is the probability that the sample belongs to class i. The entropy calculation for multiple features is calculated as in Eq. (3). is the current feature and X is the selected feature. Information gain is a measure of how well a particular feature discriminates against target classifications (Yuvaraj et al., 2021). Information gain and entropy are inversely proportional. The information gain is calculated based on the given feature values, the difference between the entropy before the split and the average entropy after the data set to split. Information gain is calculated as in Eq. (4). These approaches calculate values for each feature. Values are sorted and features are placed in the tree in order. In other words, the feature with the highest information gain and the lowest entropy is placed at the root.

Random forest

Random Forest (RF) methods, one classifier instead of multiple generating many classifiers and then from their predictions learning that classifies new data with the votes received are algorithms. The RF method is considered as collective learning. In other words, more accurate results are obtained because it finds the average by using the results of more than one decision tree. When performing Random Forests based on classification data, the Gini index is usually calculated. GINI index measures class homogeneity. It is used to decide how the nodes in the Decision Tree branch are knotted. The Gini index formula is given in Eq. (5). Here, i represents the probability that the sample belongs to the class, and c represents the number of classes (Yang, Liu, Liu, & Li, 2021).

Naive Bayes

The Naive Bayes (NB) method considers all the features to be equally important and independent from each other. It is based on Bayes’ Theorem. In the NB method, . X is an example from a test data set with m features. Let , ,…, be possible class values (Jahanbani Fard, 2015). In Eq. (6), the probabilities of the features according to the classes are calculated. In Eq. (7), the probability of each class is calculated to find out which class the data sample belongs to. The class with the highest result is designated as the class of the data sample.

Experimental results

Recently, deep learning and traditional machine learning have become very popular in many disease diagnostic applications using medical images. Deep learning models have been developed for the diagnosis of COVID-19 using both CT and X-ray images (Ardakani et al., 2020, Oh et al., 2020, Singh et al., 2020, Ucar and Korkmaz, 2020, Wu et al., 2020). Traditional machine learning methods are being used actively for COVID-19 detection, prognosis and epidemic prediction and has contributed to greatly reduce the severity of the epidemic (Islam et al., 2021). In this study, the most commonly used classification methods in the context of disease diagnosis via medical images and eight pre-trained deep learning models were used to distinguish COVID-19 positive cases from the COVID-19 negative group. The study was carried out on MATLAB R2021a on a computer with Intel(R) Core (TM) i7-8550U CPU @ 1,80 GHz.16 GB RAM. 20% of the data set was used for testing. Performance Metrics were calculated by creating a confusion matrix. Accuracy, sensitivity, specificity, precision and F1-score are calculated using equations 8, 9, 10, 11, 12 (Bozkurt, 2021). TP: The image is estimated to be COVID-19, and this image is indeed COVID-19. TN: The image is estimated to be normal and this image is normal. FP: The image is estimated to be COVID-19, but this image is normal. FN: The image is estimated to be normal, but this image is COVID-19 (Afify & Zanaty, 2021). The ROC curve is plotted by comparing the false positive rate (1-Specificity) with the correct positive rate (Sensitivity) at certain threshold settings. The area covered by the ROC curve gives the AUC value. It tells how much the model can distinguish between classes.

Experimental results of ResNet-101

The deep features obtained with the ResNet-101 deep learning model were used in classification methods such as SVM, kNN, RF, DT and NB, and the results were compared. The results of the classification methods are reported in detail in Table 4. Accordingly, by looking at the specificity value, it is clear that SVM and RF methods are more successful than other methods in detecting negative images of COVID-19. In addition, since the specificity value is higher than the sensitivity value in all classification methods except NB, it can be said that negative image detection is more successful than positive image detection.
Table 4

Experimental results of ResNet-101.

Accuracy (%)Sensitivity (%)Specificity (%)F1L-score (%)Precision (%)
NB82.96383.60782.43281.679.688
SVM90.3786.06693.91988.98392.105
kNN84.81583.60785.81183.26682.927
DT76.29670.49281.08175.43972.882
RF91.18590.16493.24390.90991.667
Fig. 2 shows the ResNet-101 ROC curve. As seen in the figure, the AUC value was 0.8975 in the NB method, 0.776 in the DT method, 0.8471 in the kNN method, 0.9432 in the SVM method and 0.9538 in the RF method. In the figure, it is seen that the classification method that best distinguishes the two classes is the RF method. Looking at the F1-score, accuracy value and ROC-curve, the Random Forest method obtained the highest values in all three success criteria.
Fig. 2

ROC-curve of ResNet-101.

Experimental results of ResNet-101. ROC-curve of ResNet-101.

Experimental results of ResNet-50

The deep features obtained with the ResNet-50 deep learning model were used in classification methods such as SVM, kNN, RF, DT and NB, and the results were examined. In Table 5, the results of the classification methods are given in detail. Accordingly, by looking at the specificity value, it is seen that SVM, RF and kNN methods are more successful in detecting negative images of COVID-19 than other methods. Since the SVM method showed very high success in detecting both negative and positive images, it achieved higher accuracy than all classification methods. Contrary to other methods, the sensitivity value of the NB and DT methods was higher than the specificity value. In other words, in these methods, COVID-19 positive images were detected better than negative images.
Table 5

Experimental results of ResNet-50.

Accuracy (%)Sensitivity (%)Specificity (%)F1-score (%)Precision (%)
NB83.33384.42682.43282.07279.845
SVM96.29695.08297.29795.86896.667
kNN90.74183.60796.62289.08395.327
DT75.18580.32870.94674.52569.504
RF93.70491.80395.2792.94694.118
Fig. 3 shows the ResNet-50 ROC-curve. As seen in Fig. 3, the AUC value was 0.8677 in the NB method, 0.7624 in the DT method, 0.9011 in the kNN method, 0.9821 in the SVM method, and 0.9656 in the RF method. In the figure, it is seen that the classification method that best distinguishes the two classes is the SVM method. Looking at the F1-score, accuracy value and ROC curve, the SVM method showed the highest values in all three success criteria.
Fig. 3

ROC-curve of ResNet-50.

Experimental results of ResNet-50. ROC-curve of ResNet-50.

Experimental results of AlexNet

The deep features obtained with the AlexNet deep learning model were used in classification methods such as SVM, kNN, RF, DT and NB, and the results were compared with each other. The results of the classification methods are reported in detail in Table 6. Accordingly, by looking at the specificity value, it is clear that SVM and RF methods are more successful than other methods in detecting negative images of COVID-19.
Table 6

Experimental results of AlexNet.

Accuracy (%)Sensitivity (%)Specificity (%)F1-score (%)Precision (%)
NB78.51979.50877.70376.98474.615
SVM89.6385.24693.24388.13691.228
kNN85.18582.78787.16283.47184.167
DT74.07481.96767.56874.07467.568
RF91.48187.70594.59590.29593.043
Fig. 4 shows the AlexNet ROC-curve. As seen in Fig. 4, the AUC value was 0.8348 in the NB method, 0.7558 in the DT method, 0.8497 in the kNN method, 0.9384 in the SVM method, and 0.9492 in the RF method. In the figure, it is seen that the classification method that best distinguishes the two classes is the RF method. Looking at the F1-score, accuracy and ROC curve, the RF method showed the highest values in all three success criteria.
Fig. 4

ROC-curve of AlexNet.

Experimental results of AlexNet. ROC-curve of AlexNet. ROC-curve of Vgg-19.

Experimental results of Vgg-19

The deep features obtained with the Vgg-19 deep learning model were used in classification methods such as SVM, kNN, RF, DT and NB, and the results were compared. The results of the classification methods are recorded in detail in Table 7. Accordingly, looking at the specificity value, it is seen that NB and RF methods are more successful in detecting negative images of COVID-19 than other methods. Considering the sensitivity value, SVM and RF methods obtained the most successful results in detecting positive images. Since the RF method achieved the highest results in both sensitivity and specificity, the accuracy value of the RF method was higher than the other methods. In addition, since the specificity value is higher than the sensitivity value in other classification methods except for SVM, it can be said that negative image detection is more successful than positive image detection.
Table 7

Experimental results of Vgg-19.

Accuracy (%)Sensitivity (%)Specificity (%)F1-score (%)Precision (%)
NB85.92681.96789.18984.03486.207
SVM85.92688.52583.78485.03981.818
kNN80.74175.4185.13577.96680.702
DT72.22267.21376.35168.61970.085
RF87.77886.06689.18986.4286.777
Fig. 5 shows the AlexNet ROC-curve. As seen in Fig. 5, the AUC value was 0.9156 in the NB method, 0.7321 in the DT method, 0.8027 in the kNN method, 0.9119 in the SVM method, and 0.9362 in the RF method. In the figure, it is seen that the classification method that best distinguishes the two classes is the RF method. Looking at the F1-score, accuracy and ROC curve, the RF method showed the highest values in all three success criteria.
Fig. 5

ROC-curve of Vgg-19.

Experimental results of Vgg-19.

Experimental results of Vgg-16

The results obtained by using deep features obtained by Vgg-16 deep learning model in classification methods such as SVM, kNN, RF, DT and NB were examined. The results of the classification methods are reported in detail in Table 8. Accordingly, by looking at the specificity and sensitivity value, it is clear that the SVM method is more successful than other methods in both detecting negative images of COVID-19 and detecting positive images. In addition, it can be said that negative image detection is more successful than positive image detection because the specificity value is higher than the sensitivity value in all classification methods.
Table 8

Experimental results of Vgg-16.

Accuracy (%)Sensitivity (%)Specificity (%)F1-score (%)Precision (%)
NB84.81579.50889.18982.55385.841
SVM91.11188.52593.2439091.525
kNN83.3373.7791.21679.99987.379
DT77.03774.45979.05474.5974.59
RF88.51985.24691.21688.88987.029
Fig. 6 shows the Vgg-16 ROC curve. As seen in Fig. 6, the AUC value was found to be 0.9288 in the NB method, 0.7871 in the DT method, 0.8249 in the kNN method, 0.9512 in the SVM method, and 0.9359 in the RF method. In the figure, it is seen that the classification method that best distinguishes the two classes is the SVM method. Looking at the F1-score, accuracy value and ROC curve, the SVM method showed the highest values in all three success criteria.
Fig. 6

ROC-curve of Vgg-16.

Experimental results of Vgg-16. ROC-curve of Vgg-16.

Experimental results of GoogLeNet

The deep features obtained with the GoogLeNet deep learning model were used in classification methods such as SVM, kNN, RF, DT and NB, and the results were compared. In Table 9, the results of the classification methods are given in detail. Accordingly, by looking at the specificity value, it is clear that kNN and RF methods are more successful than other methods in detecting negative images of COVID-19. The sensitivity value of the NB and DT methods was higher than the specificity value. In other words, in these methods, COVID-19 positive images were detected better than negative images.
Table 9

Experimental results of GoogLeNet.

Accuracy (%)Sensitivity (%)Specificity (%)F1-score (%)Precision (%)
NB61.85277.86948.64964.84755.556
SVM81.11181.14881.08179.51877.953
kNN79.6372.95185.13576.39580.18
DT71.85272.13171.62269.84167.692
RF84.07480.32887.16282.00983.761
Fig. 7 shows the GoogLeNet ROC curve. As seen in Fig. 7, the AUC value was 0.5994 in the NB method, 0.728 in the DT method, 0.7904 in the kNN method, 0.8405 in the SVM method, and 0.898 in the RF method. In the figure, it is seen that the classification method that best distinguishes the two classes is the RF method. Looking at the F1-score, accuracy and ROC curve, the RF method showed the highest values in all three success criteria.
Fig. 7

ROC-curve of GoogLeNet.

Experimental results of GoogLeNet. ROC-curve of GoogLeNet.

Experimental results of SqueezeNet

The deep features obtained with the SqueezeNet deep learning model were used in classification methods such as SVM, kNN, RF, DT and NB, and the results were examined. The results of the classification methods are reported in detail in Table 10. Accordingly, it is clear that RF, SVM and kNN methods are successful in detecting negative images of COVID-19 by looking at the specificity value. Considering the sensitivity value, it was seen that the SVM method was more successful in detecting COVID-19 positive images than other methods. In addition, negative image detection was more successful than positive image detection in all classification methods.
Table 10

Experimental results of SqueezeNet.

Accuracy (%)Sensitivity (%)Specificity (%)F1-score (%)Precision (%)
NB8067.21390.54175.22985.417
SVM89.6387.70591.21688.4389.167
kNN78.88963.93491.21673.23985.714
DT71.18565.57477.02767.79770.175
RF88.88980.32895.94686.72694.231
Fig. 8 shows the SqueezeNet ROC curve. As seen in Fig. 8, the AUC value was 0.897 in the NB method, 0.7198 in the DT method, 0.7758 in the kNN method, 0.9415 in the SVM method, and 0.9516 in the RF method. In the figure, it is seen that the classification method that best distinguishes the two classes is the RF method. Looking at the F1-score and accuracy value, the SVM method showed the highest values, while the ROC curve showed the highest value in the RF method.
Fig. 8

ROC-curve of SqueezeNet.

Experimental results of SqueezeNet. ROC-curve of SqueezeNet.

Experimental results of Xception

The deep features obtained with the Xception deep learning model were used in classification methods such as SVM, kNN, RF, DT and NB, and the results were compared. The results of the classification methods are reported in detail in Table 11. Accordingly, by looking at the specificity value, it is clear that the SVM method is more successful than other methods in detecting negative images of COVID-19. Considering the sensitivity value, it is seen that the RF method is more successful in detecting COVID-19 positive images than other methods. In addition, since the specificity value is higher than the sensitivity value in other classification methods except for DT and RF methods, it can be said that negative image detection is more successful than positive image detection.
Table 11

Experimental results of Xception.

Accuracy (%)Sensitivity (%)Specificity (%)F1-score (%)Precision (%)
NB83.70483.60783.78482.25880.952
SVM89.6386.88591.89288.33389.831
kNN87.03785.24688.51485.59785.95
DT76.29677.04975.67674.60372.308
RF88.88991.80386.48688.18984.848
The Xception ROC curve is given in Fig. 9. As seen in Fig. 9, the AUC value was 0.8946 in the NB method, 0.8001 in the DT method, 0.8688 in the kNN method, 0.9425 in the SVM method, and 0.9396 in the RF method. In the figure, it is seen that the classification method that best distinguishes the two classes is the SVM method. Looking at the F1-score, accuracy value and ROC curve, the SVM method showed the highest values in all three success criteria.
Fig. 9

ROC-curve of Xception.

Experimental results of Xception. In general, as seen in Table 12, the highest performance was obtained with the ResNet-50 and SVM methods. The accuracy value is 96.296%, the F1-score value is 95.858, and the AUC value is 0.9821. Looking at the runtime (RT), it is seen that ResNet-50 works faster than the other four models (ResNet-101, Xception, Vgg19 and Vgg-16). The lowest performance was obtained with GoogLeNet and NB method. The accuracy value was 61.852%, the F1-score value was 64.847, the AUC value was 0.5994, and the run time was 81.303 s Looking at the RT value in Table 12, the accuracy value of the fastest (22.58 s) AlexNet and kNN method is 85.249%. The slowest running (528.84 s) Vgg-19 and the accuracy value of the RF method is 87.778%. It is clear that AlexNet and SqueezeNet models are faster than other models. However, the accuracy values of SqueezeNet and GoogLeNet models are not high. The Vgg-19 is both slow and the accuracy value is lower than other models. In this case, it can be said that the best choice is ResNet-50. It can be said that the run time is reduced due to the small number of layers of the models and the 1 × 1 size filters. 1 × 1 filtering reduces size at depth. The pooling layer increases the number of channels while reducing the height and width of the input, that is, the depth is increased. As the depth increases, the run time increases. Except for the GoogLeNet model, the lowest performance was obtained with the DT method in other models.
Table 12

Experimental results.

Accuracy (%)Sensitivity (%)Specificity (%)F1-score (%)Precision (%)AUCRT (s)
ALEXNETNB78.51979.50877.70376.98479.5080.834822.672
SVM89.6385.24693.24388.13685.2460.938449.124
kNN85.18582.78787.16283.47182.7870.849722.582
DT74.07481.96767.56874.07481.9670.755823.181
RF91.48187.70594.59590.29587.7050.949271.914

RESNET50NB83.33384.42682.43282.07284.4260.8677138.44
SVM96.29695.08297.29795.86895.0820.9821159.39
kNN90.74183.60796.62289.08383.6070.9011122.76
DT75.18580.32870.94674.52580.3280.7624133.78
RF93.70491.80395.2792.94691.8030.9656145.85

RESNET101NB82.96383.60782.43281.683.6070.8975229.34
SVM90.3786.06693.91988.98386.0660.9432260.63
kNN84.81583.60785.81183.26683.6070.8471220.09
DT76.29670.49281.08175.43970.4920.776220.51
RF91.18590.16493.24390.90990.1640.9538247.92

VGG16NB84.81579.50889.18982.55379.5080.9288362.52
SVM91.11188.52593.2439088.5250.9512414.68
kNN83.3373.7791.21679.99973.770.8249419.68
DT77.03774.45979.05474.5974.4590.7871380.75
RF88.51985.24691.21688.88985.2460.9359387.55

VGG19NB85.92681.96789.18984.03481.9670.9156429.28
SVM85.92688.52583.78485.03988.5250.9119518.31
kNN80.74175.4185.13577.96675.410.8027482.86
DT72.22267.21376.35168.61967.2130.7321451.48
RF87.77886.06689.18986.4286.0660.9362528.84

SQUEEZENETNB8067.21390.54175.22967.2130.89741.049
SVM89.6387.70591.21688.4387.7050.941556.241
kNN78.88963.93491.21673.23963.9340.775856.330
DT71.18565.57477.02767.79765.5740.719842.535
RF88.88980.32895.94686.72680.3280.951651.164

XCEPTIONNB83.70483.60783.78482.25883.6070.8946339.08
SVM89.6386.88591.89288.33386.8850.9425416.96
kNN87.03785.24688.51485.59785.2460.8688337.14
DT76.29677.04975.67674.60377.0490.8001356.01
RF88.88991.80386.48688.18991.8030.9396373.29

GOOGLENETNB61.85277.86948.64964.84777.8690.599481.303
SVM81.11181.14881.08179.51881.1480.840586.217
kNN79.6372.95185.13576.39572.9510.790483.822
DT71.85272.13171.62269.84172.1310.728232.09
RF84.07480.32887.16282.00980.3280.89898.620
ROC-curve of Xception. As the number of epochs increases, the performance of the model increases significantly. Training can be terminated at these points, as performance will increase in very small units after a certain epoch. In this study, the training was terminated at the 6th epoch, as the success would increase very little after the 6th epoch in our trials. Hyper parameters were selected by trial and error method. By iteratively changing the hyperparameters one after the other, the success of the model was observed and the most suitable hyper parameter group was tried to be selected for the model. Fig. 10 shows the training and testing process of the ResNet-50 model. As can be seen in the figure, training was conducted for a maximum of 6 epochs. An epoch is a complete training cycle on the entire training dataset. The fc layer, softmax layer and output layer are changed and transferred to the new classification task. In these layers, the learning rate is set to 20 and learning is accelerated in the last layers. The initial learning rate in the transfer layers was determined as 0.0001. Thus, learning is slowed down in these layers. The verification frequency is set to 10 iterations.
Fig. 10

The Training and Testing Process of ResNet-50.

Experimental results. As can be seen in Table 13, using deep learning models together with classification methods, instead of using only deep learning model (ResNet-50), reduces the run time. While classifying with ResNet-50, 7458 s run time is calculated. When ResNet-50 and SVM methods classify together, 159.39 s run time is calculated. A deep learning algorithm takes a long time to train because there are so many parameters. Traditional machine learning takes relatively little time, ranging from a few minutes to several hours. Traditional machine learning systems can be set up and run quickly, but these systems require structured data. In complex problems, the power of its results may be limited. Deep learning systems take more time to set up, but using deep learning for complex computations (unstructured data) will strengthen results. Therefore, in this study, it can be said that by creating a hybrid model, extracting features with deep learning models and using them in traditional machine learning methods shortens the time and strengthens the results. ResNet-50 alone detected the COVID-19 images with 94.262% accuracy, but when ResNet-50 and SVM method were used together, it detected the COVID-19 images with 95.082% accuracy.
Table 13

Classification experimental results with ResNet-50.

Accuracy (%)Sensitivity (%)Specificity (%)F1-score (%)Precision (%)RT (s)
92.5994.26291.21691.99989.8447458
The Training and Testing Process of ResNet-50. Classification experimental results with ResNet-50.

Discussion

We combined classification methods (traditional machine learning) and capabilities extracted from deep learning models to increase the accuracy value in COVID-19 detection. In this review, we obtained deep features for COVID-19 detection using 8 powerful deep learning models on CT images. Comparative analysis was performed using 5 different classification methods for each deep feature. On evaluating We concluded that using the deep learning model and classification method together provides greater accuracy in the evaluation. Using deep learning models together with classification methods, instead of using only deep learning model (ResNet-50), reduces the run time. While classifying with ResNet-50, 7458 s run time is calculated. When ResNet-50 and SVM methods classify together, 159.39 s run time is calculated. ResNet-50 alone detected the COVID-19 positive images with 94.262% accuracy, but when ResNet-50 and SVM method were used together, it detected the COVID-19 images with 95.082% accuracy. Alternatively, using ResNet-50 and SVM together, it achieved 96.296% COVID-19 detection accuracy for the dataset. The purpose of this study is to reduce the duration and amount of transmission of the disease by shortening the diagnosis time by taking CT images from patients who were admitted to the hospital due to the suspicion of COVID-19. In addition, it is aimed to provide a decision support system to radiologists in the detection of COVID-19. The COVID-19 pandemic is a new disease problem, and there is no definite and sufficient data in the literature on this subject. All studies except (Wu et al., 2021) study in the literature could not reach the COVID-19 patient image as in this study. The data set used in this study consists of real images obtained from the hospital environment. In addition, images thought to be positive for COVID-19 in this study were confirmed by a radiologist. The working times of the methods used in most of the studies in the literature are not mentioned. In this study, methods were evaluated according to accuracy, sensitivity, specificity, precision, F1-score, AUC and run time. Compared to the studies in the literature, a broad evaluation was made in this study using various classification methods together with deep learning models. The performance evaluation of the proposed model compared to studies using CT images in the literature is given in Table 14. Since it is more important to accurately detect COVID-19 positive cases, sensitivity values are taken into account in Table 14. The sensitivity value of the proposed model (95.082%) achieved higher performance than all models based on CT images in the studies in the literature.
Table 14

Comparison of the proposed model with studies using CT images.

StudyMethodNumber of imagesData sourcePerformance
Wu et al. (2020)COVNet4356Multiple hospital environment90% sensitivity
Zheng et al. (2020)DeCOVNet630Multiple hospital environment90.7% sensitivity
Singh et al. (2020)MODE-CNN150Open source database90.7% sensitivity
Pham (2020)DenseNet-201746Open source database96.2% accuracy
Wu et al. (2021)COVIDAL962Open source database86.6% accuracy
Rohila et al. (2021)ReCOV-111110MosMedData94.9% accuracy
Wang, Kang et al. (2021)Modified Inception V31065Multiple hospital environment85.2% accuracy
Li et al. (2021)CheXNet746Open source database87% accuracy

This studyResNet-50 - SVM1345Single hospital environment96.96% accuracy 95.082% sensitivity
Pham (2020) examined transfer learning (pretrained convolutional neural networks) methods in his study. In our study, various hybrid models were created by using the deep features obtained from different deep learning methods in different machine learning methods for the detection of COVID-19. The 1345 CT images in the dataset used were obtained from the hospital environment. It consists of various CT images (COVID-19 early stage or later stages, different ages and different genders). The performances of the new hybrid models created were examined. In the results of the experiment, it was seen that more successful results were obtained by using the hybrid model created by integrating deep learning and machine learning methods, as in our study, instead of using deep learning models alone as in the Pham (2020) study. Comparison of the proposed model with studies using CT images. The COVID-19 pandemic is a new disease problem, and there is no definite and sufficient data in the literature on this subject. All studies except (Wu et al., 2020) study in the literature could not reach the COVID-19 patient image as in this study. The data set collected consists of real images obtained from the single hospital environment. Images thought to be positive for COVID-19 in this study were confirmed by a radiologist. Most of the datasets used were created from open-source databases. The data set in some studies in the literature (Mahmud et al., 2020, Oh et al., 2020, Ucar and Korkmaz, 2020) contains a large number of images, but the number of images of COVID-19 cases is low. In addition, some studies (Panwar et al., 2020, Singh et al., 2020) have datasets containing a small number of images. In both cases, a limited number of data leads to underfitting or overfitting, reducing the performance of deep learning models (Islam et al., 2021, Nayak et al., 2021). In some of the studies in the literature (Mahmud et al., 2020, Oh et al., 2020, Ucar and Korkmaz, 2020) data imbalance is observed. Data imbalance also increases bias to a class (Islam et al., 2021). The run times of the methods used in most of the studies in the literature are not mentioned.

Conclusion

Since the patient is not quarantined during the conclusion of the Polymerase Chain Reaction (PCR) test used in the diagnosis of COVID-19, the disease continues to spread. In this study, it was aimed to reduce the duration and amount of transmission of the disease by shortening the diagnosis time of COVID-19 patients with the use of Computed Tomography (CT). In addition, it is aimed to provide a decision support system to radiologists in the diagnosis of COVID-19. In this study, a new high-performance COVID-19 detection system was developed with the hybrid model created from the combination of ResNet-50 and SVM. In our proposed hybrid model framework, there are 2 modules including feature extraction with deep learning method and classification with traditional machine learning method. The data set used in this study consists of real images obtained from a single hospital environment. Images thought to be positive for COVID-19 were confirmed by a radiologist. According to the experimental results, the proposed hybrid model (ResNet-50SVM) achieved higher accuracy than the classical ResNet-50 method. In the future, images from more than one hospital can be collected and better results can be obtained. Different lung diseases can be added in future studies.
  23 in total

1.  Harnessing the Power of Smart and Connected Health to Tackle COVID-19: IoT, AI, Robotics, and Blockchain for a Better World.

Authors:  Farshad Firouzi; Bahar Farahani; Mahmoud Daneshmand; Kathy Grise; Jaeseung Song; Roberto Saracco; Lucy Lu Wang; Kyle Lo; Plamen Angelov; Eduardo Soares; Po-Shen Loh; Zeynab Talebpour; Reza Moradi; Mohsen Goodarzi; Haleh Ashraf; Mohammad Talebpour; Alireza Talebpour; Luca Romeo; Rupam Das; Hadi Heidari; Dana Pasquale; James Moody; Chris Woods; Erich S Huang; Payam Barnaghi; Majid Sarrafzadeh; Ron Li; Kristen L Beck; Olexandr Isayev; Nakmyoung Sung; Alan Luo
Journal:  IEEE Internet Things J       Date:  2021-04-19       Impact factor: 10.238

2.  Sensitivity of Chest CT for COVID-19: Comparison to RT-PCR.

Authors:  Yicheng Fang; Huangqi Zhang; Jicheng Xie; Minjie Lin; Lingjun Ying; Peipei Pang; Wenbin Ji
Journal:  Radiology       Date:  2020-02-19       Impact factor: 11.105

3.  Deep learning-based multi-view fusion model for screening 2019 novel coronavirus pneumonia: A multicentre study.

Authors:  Xiangjun Wu; Hui Hui; Meng Niu; Liang Li; Li Wang; Bingxi He; Xin Yang; Li Li; Hongjun Li; Jie Tian; Yunfei Zha
Journal:  Eur J Radiol       Date:  2020-05-05       Impact factor: 3.528

4.  Application of deep learning technique to manage COVID-19 in routine clinical practice using CT images: Results of 10 convolutional neural networks.

Authors:  Ali Abbasian Ardakani; Alireza Rajabzadeh Kanafi; U Rajendra Acharya; Nazanin Khadem; Afshin Mohammadi
Journal:  Comput Biol Med       Date:  2020-04-30       Impact factor: 4.589

5.  A comprehensive study on classification of COVID-19 on computed tomography with pretrained convolutional neural networks.

Authors:  Tuan D Pham
Journal:  Sci Rep       Date:  2020-10-09       Impact factor: 4.379

6.  A Unified Hierarchical XGBoost model for classifying priorities for COVID-19 vaccination campaign.

Authors:  Luca Romeo; Emanuele Frontoni
Journal:  Pattern Recognit       Date:  2021-07-22       Impact factor: 7.740

7.  Application of deep learning for fast detection of COVID-19 in X-Rays using nCOVnet.

Authors:  Harsh Panwar; P K Gupta; Mohammad Khubeb Siddiqui; Ruben Morales-Menendez; Vaishnavi Singh
Journal:  Chaos Solitons Fractals       Date:  2020-05-28       Impact factor: 5.944

8.  Recommendation of low-dose CT in the detection and management of COVID-2019.

Authors:  Zhen Kang; Xu Li; Shuchang Zhou
Journal:  Eur Radiol       Date:  2020-03-19       Impact factor: 5.315

9.  A deep learning algorithm using CT images to screen for Corona virus disease (COVID-19).

Authors:  Shuai Wang; Bo Kang; Jinlu Ma; Xianjun Zeng; Mingming Xiao; Jia Guo; Mengjiao Cai; Jingyi Yang; Yaodong Li; Xiangfei Meng; Bo Xu
Journal:  Eur Radiol       Date:  2021-02-24       Impact factor: 5.315

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.