Literature DB >> 35935665

CoviDetNet: A new COVID-19 diagnostic system based on deep features of chest x-ray.

Muzaffer Aslan1.   

Abstract

COVID-19 has emerged as a global pandemic affecting the world, and its adverse effects on society still continue. So far, about 243.57 million people have been diagnosed with COVID-19, of which about 4.94 million have died. In this study, a new model, called COVIDetNet, is proposed for automated COVID-19 detection. A lightweight CNN architecture trained instead of the popular and pretrained convolution neural network (CNN) models such as VGG16, VGG19, AlexNet, ResNet50, ResNet100, and MobileNetV2 from scratch with chest x-ray (CXR) images was designed. A new feature set was created by concatenating the features of all layers of the designed CNN architecture. Then, the most efficient features chosen among the features concatenating with the Relief feature selection algorithm were classified using the support vector machine (SVM) method. The experimental works were carried out on a public COVID-19 CXR database. Experimental results demonstrated 99.24% accuracy, 99.60% specificity, 99.39% sensitivity, 99.04% precision, and an F1 score of 99.21%. Also, in comparison to AlexNet and VGG16 models, the deep feature extraction durations were reduced by approximately 6-fold and 38-fold, respectively. The COVIDetNet model provided a higher accuracy score than state-of-the-art models when compared to multi-class research studies. Overall, the proposed model will be beneficial for specialist medical staff to detect COVID-19 cases, as it provides faster and higher accuracy than existing CXR-based approaches.
© 2022 Wiley Periodicals LLC.

Entities:  

Keywords:  COVID‐19; Relief; SVM; automatic detection; deep feature extraction with a lightweight CNN

Year:  2022        PMID: 35935665      PMCID: PMC9347592          DOI: 10.1002/ima.22771

Source DB:  PubMed          Journal:  Int J Imaging Syst Technol        ISSN: 0899-9457            Impact factor:   2.177


INTRODUCTION

Cases of pneumonia of unknown cause were detected in the city of Hubei, Wuhan Province, China, in December 2019. In January 2020, the World Health Organization (WHO) declared these cases to be due to a new type of virus, and in February 2020, they named this virus COVID‐19. The virus's rapid spread among people has become a severe public health problem worldwide. Health systems have come to a standstill, especially in developed countries, due to rapidly rising cases. Many countries have been forced to take social isolation measures to maintain health services and reduce the risk of viral transmission. Additionally, many countries have banned their citizens from traveling, both home and abroad. Despite all these precautions, 251.26 million cases have been detected worldwide, according to WHO data dated November 12, 2021. So far, 5.07 million people have died due to COVID‐19 infection. The highest confirmed cases and deaths were found in the Americas during this period, followed by Europe. COVID‐19 has caused health problems in people and caused severe damage to people's economic and social lives. This situation mainly affected developed and developing countries such as the United States, India, the United Kingdom, Brazil, and the Russian Federation, where COVID‐19 cases and deaths are very high. Figure 1 indicates that the number of cases and deaths in the countries most affected by COVID‐19 in the world, according to WHO statistics.
FIGURE 1

The total numbers of cases and deaths reported by World Health Organization from different countries are shown in the chart. (The columns and line show the total number of cases and total deaths affected by COVID‐19 from January 3, 2020, to November 12, 2021, respectively.)

The total numbers of cases and deaths reported by World Health Organization from different countries are shown in the chart. (The columns and line show the total number of cases and total deaths affected by COVID‐19 from January 3, 2020, to November 12, 2021, respectively.) As can be shown in the graph in Figure 2, until November 12, 2021, in the United States, India, Brazil, United Kingdom, Russian Federation, Turkey, and France, 46.41, 34.40, 21.89, 9.4, 8.95, 8.31, and 7.01 million, respectively, cases have been identified. Also, in Figure 2, the United States, India, Brazil, United Kingdom, Russian Federation, Turkey, and France are 751.60, 462.19, 609.75, 142.33, 251.69, 72.71, and 115.61 a thousand, respectively, deaths have been reported. Despite vaccination processes, which started at the beginning of 2021, the epidemic continues to spread rapidly due to the mutational effects of the virus. Some experts indicate that the epidemic's effects may last until 2024.
FIGURE 2

Structure of the proposed COVIDetNet model

Structure of the proposed COVIDetNet model Some of the widespread symptoms of the early periods of COVID‐19 infection are also seen in people affected by colds and other flu virus infections. Therefore, disease diagnostic test kits are needed for the definitive diagnosis of COVID‐19. Real‐time reverse transcription‐polymerase chain reaction (RT‐PCR) testing on respiratory or blood samples are some of the effective methods of choice for COVID‐19 diagnosis. However, these kits have disadvantages, such as long diagnosis times (1–3 days), costly, and sometimes, failure to detect the virus. Some studies have stated that only 71% of COVID‐19 cases can be detected with RT‐PCR. Recently, medical imaging data such as x‐rays, computed tomography (CT), and magnetic resonance imaging have been used to detect and diagnose many diseases such as brain tumors and lung. , For this reason, lung radiological chest scans are used in COVID‐19 case detection. CT and chest x‐ray (CXR) are used for breast scanning. Radiologists and physicians can distinguish positive COVID‐19 cases with the help of radiological lung imaging; this method is preferred for diagnosis as it gives both faster and more precise results when compared to the RT‐PCR method. CT scanning systems are new and expensive. Therefore, they are not available in all medical centers and hospitals. Meanwhile, CXR systems are found in almost all medical centers and hospitals, and, in addition to being light and portable, these systems provide an excellent convenience in their use. An average of 15–25 s may be sufficient for each patient for the CXR procedure. Therefore, radiologists and physicians prefer CXR images, as they are more economical and faster for case diagnosis. However, radiologists need time to examine CXR images. As a result, developing rapid and highly accurate automatic case detection systems based on artificial intelligence that can reduce radiologists' workload is critical.

Literature review

Narin et al. proposed the ResNet50, Inception V3, and Inception‐ResNetV2 models for diagnosing COVID‐19 from CXR images. In their studies, the ResNet50 model was shown to perform better than the other two models, with 98.0% classification accuracy. Ismael and Şengür proposed a combination of pretrained convolutional neural network (CNN) ResNet, VGG models, and an support vector machine (SVM) classifier to diagnose COVID‐19. In their study, feature extraction was initially performed using CXR images with ResNet and VGG models; then, the SVM classified suspected COVID‐19 cases as either positive or negative, demonstrating 94.7% accuracy. Hemdan et al. proposed a deep learning (DL) model named COVIDX‐Net; the method they proposed achieved 90% success in testing of 50 CXR images. Similarly, Khalifa et al. proposed a generative adversarial network, based on DL, to diagnose COVID‐19 cases using limited CXRs. Thus, CNN models, such as ResNet18, AlexNet, GoogLeNet, and SqueezeNet, solved overfitting and achieved higher performance. Chowdhury et al. proposed eight different CNN models, including AlexNet, DenseNet201, ResNet18, SqueezeNet, and MobileNet, pretrained to detect COVID‐19 and pneumonia. Their work combined various publicly available datasets to create a dataset containing 423 COVID‐19 positive cases, 1485 pneumonia cases, and 1579 normal CXR images; in their experimental studies, they achieved a 97.74% accuracy score in triple classification. Hussain et al. proposed a 22‐layer CNN architecture for COVID‐19 detection from CXR images. The proposed CNN model features were classified into 2, 3, and 4 classes. Their study achieved accuracy scores of 99.1%, 94.2%, and 91.2%, respectively. Ibrahim et al. suggested pretrained CNN models to detect COVID‐19 from CXR and CT images. The referenced study used CXR images to detect normal physiology, COVID‐19 cases, and pneumonia, while also using CT images to detect normal, COVID‐19, pneumonia, and lung cancer cases. For this purpose, they used four different CNN models: ResNet, ResNet + Bidirectional, ResNet + Gated Recurrent Unit (ResNet + GRU), and VGG19; their researcher achieved the highest COVID‐19 detection rate using the VGG model, which demonstrated a 98.05% accuracy score in the experimental studies. Ozturk et al. suggested a CNN architecture with 17 convolution layers to diagnose COVID‐19. The researchers obtained an accuracy rate of 87.02% in studies conducted with a three‐class dataset. Turkoglu proposed a model, which they called COVIDetectioNet, in order to diagnose COVID‐19 from CXR images. The experimental works used a dataset containing 6092 CXR images labeled as either normal, COVID‐19, or pneumonia. The proposed COVIDetectioNet combined attributes from the fully connected, and the convolution layers of the AlexNet. The obtained features were classified with SVM by selecting the Relief feature selection algorithm. Their research achieved an accuracy score of 99.18% in the experimental studies. Reference 25, VGG‐16, VGG‐19, ResNet50, and Inception V3 CNN models have tried different learning rates to eliminate the effects of the overfitting problem. Also, for a faster classification, the use of the two‐layer CNN model in the transfer learning method has been proposed. They achieved an accuracy score of 90.45% in experimental studies. Apostolopoulos and Mpesiana proposed five different CNN models for COVID‐19 detection: MobileNet V2, Xception, Inception, ResNetV2, and VGG19. The research team achieved an accuracy rate of 94.72% using the MobileNetV2 architecture. Ucar et al. proposed a pretrained SqueezeNet architecture with a Bayes optimization algorithm to detect COVID‐19 and achieved a 98.26% accuracy rate in their experimental works. Wang et al. proposed a CNN model that includes many convolution layers. Using this method, they detected COVID‐19 cases with 92.64% accuracy in the proposed CNN model. Similarly, Nour et al. detected COVID‐19 with an end‐to‐end trained CNN model and obtained an accuracy score of 98.97%.

Motivation and contributions

The literature research shows promising results obtained in CNN‐based studies with CXR images. , , , , , , , , , , , , , In these studies, pretrained CNN models that were generally based on transfer learning were used. Additionally, many researchers have focused on accuracy scores and the dual classification of COVID‐19 cases as either positive or negative. However, during the pandemic period, in which case numbers are very high, it is essential how long it takes to decide on this diagnosis as well as the correct diagnosis of COVID‐19. This research study, a new CNN architecture, named COVIDetNet, is suggested to diagnose COVID‐19 automatically. The structure of the suggested COVIDetNet model is represented in Figure 2. COVIDetNet comprises three stages: deep feature extraction, feature selection, and classification. The proposed CNN model with end‐to‐end architecture was created and trained using CXR images. In this method, deep features are taken and concatenated from each convolution and the fully connected layer of the constructed CNN model. The Relief feature selection algorithms were used to select the most effective features from a set of combined features. The SVM algorithm was then used to categorize these efficient features. Based on the results of the experiments, it is clear that the method will be useful in detecting COVID‐19. The major contributions of the proposed method may be outlined as follows:The limitations of the proposed method can be explained as follows: A flexible new CNN architecture was proposed, where all layer structures and dimensions can be changed easily. The proposed CNN model requires less training time as it is a lightweight model compared to pretrained CNN models. Deep features are mostly extracted from one or some layers of deep learning models. In COVIDetNet, deep features from all convolutional and fully connected layers were obtained, making CX images' characteristic details more prominent. The Relief feature selection algorithm improved the classification performance. Besides, the size of the feature set was reduced with this algorithm. Thus, the prediction time of the classifier was reduced. Compared to popular deep learning models, the computational cost of the proposed model is very low. In this regard, the COVIDetNet approach can be used in clinical applications with low‐capacity hardware. The proposed approach outperformed existing methods in the literature (see Table 6),
TABLE 6

Performance results of state‐of‐the‐art approaches (%)

Research works/yearMethodType/number of classNumber of casesResults
Acc (%)Sen (%)Spe (%)Pre (%)FITime duration (s)
Ozturk et al. 23 /2020DarkCovidNetCXR 2

COVID‐19: 125

No findings: 500

98.0895.1395.3098.030.9651
CXR 3

COVID‐19: 125

Pneumonia: 500

No findings: 500

87.0285.3592.1889.960.8737
Turkoglu 24 /2020COVIDetectioNetCXR 3

COVID‐19: 219

Pneumonia: 4290

Normal: 1583

99.1899.1399.480.9930
Ucar et al. 27 /2020COVIDiagnosis‐NetCXR 3

COVID‐19: 1536

Pneumonia: 1536

Normal: 1536

98.2699.130.9825
Wang et al. 28 /2020COVID‐NetCXR 3

COVID‐19: 13870

Pneumonia: 5538

Normal: 8066

93.3
Nour et al. 29 /2020A novel CNN model + SVMCXR 3

COVID‐19: 219

Pneumonia: 1345

Normal: 1345

98.9789.3999.750.9672
ProposedCOVIDetNetCXR 3

COVID‐19: 500

Pneumonia: 580

Normal: 1541

99.24 99.39 99.60 99.040.9921287
Hussain et al. 21 /2021CoroDetCXR 2

COVID‐19: 500

Normal: 800

99.10
CXR 3

COVID‐19: 500

Pneumonia: 800

Normal: 800

94.20
Cengil and Çınar 11 /2021AlexNet + NASNetLargeCXR 3

COVID‐19: 1525

Pneumonia: 1525

No Findings:1525

96.0098.1098.5098.103791
NASNetLarge + XceptionCXR 3

COVID‐19: 1525

Pneumonia: 1525

No Findings:1525

97.6099.0099.2099.004485
Murugan and Goel 54 /2021E‐DiCoNetCXR 3

COVID‐19: 900

Pneumonia: 900

Normal: 900

94.0798.1591.4898.150.9122
Barua et al 55 /2021COVID‐19FclNet9CXR 3

COVID‐19: 125

Pneumonia: 500

Control: 500

89.96
CXR 3

COVID‐19: 3616

Pneumonia: 1345

Control: 4000

98.8498.760.9871
ProposedCOVIDetNetCXR 3

COVID‐19: 580

Pneumonia: 500

Normal: 1541

99.24 99.39 99.60 99.040.9921287

Note: Bold values shows the performance results where the proposed method is superior to the existing methods.

Preprocessing x‐ray images and increasing the input image size in the training phase will increase the computational cost. Therefore, powerful hardware will be required.

Organization of paper

The research is structured as follows: Section 2 gives the methodology background, consisting of the datasets, CNN architecture, and classification method. The experimental results give in Section 3. Sections 4 and 5 present the discussion and conclusion of the paper, respectively.

MATERIALS AND METHODS

In this paper, an effective and new COVIDetNet model is proposed to detect COVID‐19 cases in minimum time and with high accuracy. The general architecture of the proposed method is presented in Figure 2. The method consisted of four basic phases: In the first phase, the CXR images were resized according to the COVIDetNet model; in the second phase, deep features were obtained and concatenated from all fully connected and convolutional layers of the CNN architecture, whose layer numbers and sizes are easily changed; following that, the most effective features from the deep features acquired with the Relief algorithm were chosen; and, finally, the SVM method was used to classify the specified features.

Dataset

The dataset was created using a total of 2621 CXR images, which were compiled using various open‐source, public data sets. Radiologists labeled CXR images as either normal, COVID‐19 positive, or pneumonia. The CXR images belonging to the dataset were organized in three different folders according to their tag classes. The normal, COVID‐19, and pneumonia folders contained 1541, 580, and 500 CXR images, respectively. Normal (healthy) and COVID‐19 positive CXR images were collected from the Kaggle websites. , The COVID‐19 positive CXR images, which radiologists took, comprising data from 200 males and 161 females over 45. Pneumonia CXR images were taken from a dataset generated by Wang et al. Figure 3 demonstrates example images of CXRs from the normal, COVID‐19, and pneumonia classes, which give in the first, second, and third columns, respectively.
FIGURE 3

Chest x‐ray image samples in the dataset

Chest x‐ray image samples in the dataset

Convolution neural network

CNN is a DL algorithm that is very popular in image classification and object detection. CNN structures can be designed as one‐dimensional (1D), two‐dimensional (2D), or three‐dimensional (3D). The 1D CNN is used for serial data processing, while 2D CNN is used for image and text recognition, and 3D CNN structures are often used for video and medical image recognition. CNNs are designed in layers, as in a typical neural network model; neurons, bias, and weights interconnect the layers. However, the CNNs differ from typical neural networks in that they include three types of layers: convolution, pooling, and fully connected layers, in order for the CNNs to form an entire network. The convolution layer is an important layer for the CNN that uses convolution operation, denoted by “*”, rather than the general matrix multiplication. , This layer's learnable parameters are made up of a series of learnable filters known as kernels. , The 2D convolution operation can be defined by Equation (1): where X is the CXR input image and K is the kernel. The K kernel has n × n dimensions. This method provides the ability to obtain lower‐dimensional features with the help of movement covering all pixels on the input image with the determined stride size. , , In brief, the convolution layer creates a feature map of the original images by determining local connections of the properties belonging to the input data that may be similar to the CXR images. The batch normalization (BN) layer is used to reduce the training process of the CNN and develop network initiating performance. , Also, the BN layer is used to minimize gradient vanishing. BN layer variables are calculated as shown in Equations ((2), (3), (4), (5)): where is mini‐batch mean, is mini‐batch variance, and is the normalized activation. When is very small, constant is used to improve the numerical operation. Balance factors d and e are learnable parameters. BN output is defined by . In DL models, the convolution layer output is transferred to a nonlinear layer to increase the CNN's fitting capability. This layer, called the activation layer, uses the nonlinear rectified linear unit (ReLU) activation function to reduce gradient explosion and gradient vanishing. In place of sigmoid and tanh activation functions, which are preferred in artificial neural networks, smooth ReLU uses CNNs. The ReLU activation function can be defined as given in Equation (6): Through the pooling layer, which keeps the kernel size of the obtained feature matrix as a result of the activation process constant, the width and height of the matrix are reduced by the maximum or average value method. , The fully connected layer (fc) attaches to the previous and subsequent layer's neurons. The first fc comprises flattened matrices carried from the previous layers. Meanwhile, the last fc determines how much a data value matches any class, and passes this information to the softmax layer for classification.

The architecture of the proposed CNN model

A simple CNN that can be used on lower‐capacity hardware, called COVIDetNet, was designed for deep feature extraction from CXR images. COVIDetNet consists of a 29‐layer structure that includes six convolutions, seven ReLUs, six normalizations, four maximum pools, three fully connected layers, and a softmax layer. The COVIDetNet architecture, whose properties such as the number of layers, kernel, stride, and padding sizes were determined experimentally, are shown in Table 1 and its architecture in Figure 4.
TABLE 1

COVIDetNet architecture

LayerFilter sizeKernel sizeStride sizeOutput size
Input image100 × 100 × 3
Convolution‐13 × 3641100 × 100 × 64
Batch normalization/ReLU/max pooling2 × 2250 × 50 × 64
Convolution‐25 × 564225 × 25 × 64
Batch normalization/ReLU/max pooling2 × 2212 × 12 × 64
Convolution‐33 × 332112 × 12 × 32
Max pooling/batch normalization2 × 226 × 6 × 32
Max pooling2 × 223 × 3 × 32
Convolution‐42 × 23222 × 2 × 32
Batch normalization/ReLU2 × 2 × 32
Convolution‐53 × 31621 × 1 × 16
Batch normalization/ReLU1 × 1 × 16
Convolution‐63 × 31621 × 1 × 16
Batch normalization/ReLU1 × 1 × 16
Fully connected/ReLU1 × 1 × 500
Fully connected/ReLU1 × 1 × 350
Fully connected1 × 1 × 3
FIGURE 4

The architecture of the proposed convolution neural network

COVIDetNet architecture The architecture of the proposed convolution neural network In the literature, fc layers of pretrained CNN models have been used for deep feature extraction. However, in the current paper, convolution layer features are also combined with the fc layers. For this, the ensemble average of the features obtained from each of the six convolution layers was taken. Suppose the visual data in a convolution layer output is W × H × C; this process can be described by Equations ((7), (8), (9)) : where W × H and C represent the size of the learned visual features and the number of channels, respectively.

Relief feature selection

The Relief algorithm, was presented by Kira and Rendell. It is designed to enable feature selection. Feature selections are depended on the k nearest neighbor algorithm ; this algorithm estimates the quality of the attributes according to the best separation between properties whose values are close to each other. For this goal, the method calculates an index of interest for each feature. Then, according to the calculated index value, it gives a positive weight value (+1 best) to attributes that are close to one another, while a negative weight value (−1 is the worst) is applied to other attributes. Thus, a multivariate Relief filter is created that sorts attribute according to their index interest level. The two‐class Relief algorithm and feature selection steps are given below: Step 1: The initial weight (IW) of each attribute (A) is selected as 0 . Step 2: The R random sample is taken from the feature set. Step 3: The distances between R and the rest of the features are calculated using Equation (9), while same class examples (hits‐Ht) and other class examples (misses‐Ms) are found. Step 4: The W[A] value is updated via Equation (10) above: where I 1 and I 2 are the selected features, while r represents the number of random training samples required to update IW. Steps 3 and 4 are repeated for each feature. Thus, using the calculated IW [A] values, the desired number of features are selected from all features.

Support vector machine

SVM is a supervised machine learning model based on statistical learning theory. The SVM method generally finds the separating hyperplane between two or more classes with the maximum margin. Suppose we have a dataset {x , y and i = 1, 2, …, M} with a class label of {−1, +1} that consists of M instances. Accordingly, the optimum hyperplane is shown in Equation (12) : where w and b symbolize the hyperplane's weight vector and the trend value, respectively. The weight vector, w, must be at a minimum to maximize the optimum hyperplane boundary. In this case, determining the optimal hyperplane requires the solution of the limited optimization problem via Equation (13) below : When the data to be classified is not linearly separable, a kernel function is used, thus providing a better separation of data. Accordingly, the optimum hyperplane can be shown by Equation (14): where, , C, and represent the mapping function, the regularization parameter, and the slack variable, respectively. Additionally, the lower the value of C, the lower the restriction and the greater the probability of plane fitting. The above Equation (14) is difficult to solve, since it is in a nonconvex form. Therefore, the optimization problem is converted to a dual form with the Lagrange multiplier's equality constraint. The Lagrange function is given by Equation (15): subject to: where, represent the Lagrange multipliers. Thus, the SVM decision function in kernel space is shown by Equation (17): in which represents kernel function.

EXPERIMENTAL WORKS

The present section includes criteria for evaluating experimental studies, experimental configurations used in learning and optimization, experiments performed, and related results.

Performance evaluation criteria

The accuracy (Acc), sensitivity (Sen), specificity (Spe), precision (Pre), and F‐score metrics are used for the performance analysis of the proposed method. These metrics were computed using Equations ((18), (19), (20), (21), (22)). by which true‐positive (TP), true‐negative (TN), false‐positive (FP), and false‐negative (FN) cases are represented.

Experimental works and results

The experimental works were implemented with MATLAB 2020b on a workstation with an Intel® Core™ i7‐4700HQ processor, 16 GB of memory, and a 4 GB graphics card. For experimental works, the dataset was resized to 100 × 100 in accordance with the COVIDetNet input. Also, the dataset was randomly divided into 80% training and 20% testing. This study selected the optimization of learnable parameters, momentum stochastic gradient descent algorithm, and a learning rate of 0.0001 for all CNN models used in experimental studies. In addition, the validation frequency was selected as 30, maximum period 10, iteration 262 per period, and mini‐batch size 8. The SVM classifier's parameters were computed using the one‐versus‐all method, and the cubic kernel function. The SVM parameter C was searched in the (10−3, 103) range, the kernel scale value was set to automatic, and the box constraint level was set to 1. The 10‐fold cross‐validation approach was also applied. In the first experiment, the end‐to‐end CNN architecture from scratch was trained with x‐ray images, and deep features were extracted. The proposed COVIDetNet architecture achieved a validation accuracy score of 98.47% after 2620 iterations in 4 min 19 s as seen in Figure 5.
FIGURE 5

Graphics of loss and accuracy values for validation and training

Graphics of loss and accuracy values for validation and training Then, using the same dataset and same training parameters, feature extraction durations of AlexNet, VGG16, VGG19, ResNet50, ResNet100, and MobileNetV2 CNN models are given in Table 2. In addition, the performance criteria result of all CNN models are detailed in Table 2.
TABLE 2

The feature extraction duration and performance results for all convolution neural networks (CNNs)

CNN modelTime (s)AccSenSpePre F1 score
AlexNet142198.0897.3198.5398.7497.98
VGG16864494.4796.8697.6393.3394.65
VGG19997092.9489.9194.5195.4891.96
ResNet50521996.7697.6298.3995.9996.73
ResNet101832597.5298.5998.9496.6497.51
MobileNetV2286596.1897.4898.2295.2396.20
Proposed 259 98.47 98.0698.91 98.76 98.40

Note: Bold values shows the performance results where the proposed method is superior to the existing methods.

The feature extraction duration and performance results for all convolution neural networks (CNNs) Note: Bold values shows the performance results where the proposed method is superior to the existing methods. According to the results shown in Table 2, the proposed CNN model performed feature extraction in a short time of 259 s. The best feature extraction times of the other known methods are seen in AlexNet and MobilNetV2 models with 1421 s and 2865 s, respectively. In addition, considering the feature extraction durations in Table 4, the proposed CNN structure is about sixfold faster than the AlexNet structure and obtained the best accuracy of 98.47%. It is also 38 fold faster than the VGG19 model, with the longest feature extraction time of 9970 s.
TABLE 4

Performance results of effective features with prelearned AlexNet deep features Relief algorithm (%)

Selected feature with ReliefAccSenSpePre F1 score
100097.5297.5298.5097.3597.44
150098.2898.4999.0597.9998.23
200098.4798,6099.1398.2698.43
250098.6698.8899.2398.3898.63
300098.6698.8899.2898.3898.63
400098.6698.8899.2898.3898.63
500098.6698.8899.2898.3898.63
600098.6698.8899.2898.3898.63
700098.6698.8899.2898.3898.63
750098.4798.7899.2098.1198.43
The confusion matrix of the four models with the quickest feature extraction durations is given in Figure 6. The confusion matrix and Table 2 results in ResNet50, proposed lightweight CNN for the COVID‐19 class, and AlexNet, our model for healthy (normal) class, respectively, have high performance. In these models, MobilNetV2 96.18%, ResNet50 96.76%, AlexNet 98.08%, and the proposed CNN 98.47% accuracy rates were obtained for all three classes.
FIGURE 6

Confusion matrix of the models with the quickest feature extraction times

Confusion matrix of the models with the quickest feature extraction times Convolution and fully connected layers were used for feature extraction with the proposed CNN (see Section 2.4). Thus, 64, 64, 32, 32, 16, 16, 500, and 350 features were extracted from Conv1, Conv2, Conv3, Conv4, Conv5, Conv6, and the fc1 and fc2 layers, respectively. Next, all of the features were concatenated. The deep features obtained were classified using SVM. The experiment's confusion matrix and receiver operating characteristic (ROC) curve are shown in Figure 7.
FIGURE 7

Confusion matrix and receiver operating characteristic curves for all features

Confusion matrix and receiver operating characteristic curves for all features As seen in Figure 7, accuracy scores of 97.40%, 99.00%, and 100.00% were obtained for the COVID‐19, normal, and pneumonia classes, respectively. The mean accuracy score achieved in all three classes was 98.90%. Efficient features were selected from 1074 features concatenated with the Relief algorithm. In the experiments with Relief, features in the range of (20, 30) were chosen, according to the one–step‐size method; this range and step size were determined experimentally. Next, the performance of each chosen feature vector was calculated with SVM. Performance results of selected features with the Relief algorithm are given in Table 3.
TABLE 3

Performance results of selected features using the Relief algorithm (%)

Selected feature with ReliefAccSenSpePre F1 score
2098.2897.9598.8398.8498.21
2198.0897.8498.7598.1998.01
2298.4798.0698.9198.7698.40
2399.0599.1199.4598.9399.01
2499.0599.1099.4599.9399.01
2599.2499.3999.6099.0499.21
2699.2499.2199.5399.2199.21
2799.0598.9299.3799.1099.01
2898.8598.6399.2298.9998.81
2998.6698.5399.1498.7098.61
3098.6698.5399.1498.7098.61
Performance results of selected features using the Relief algorithm (%) In Table 3, the F scores and precision, specificity, sensitivity, and accuracy scores of the features chosen by the Relief algorithm are detailed. The Relief algorithm's maximum accuracy score was 99.24%, which was attained with 25 and 26 Relief features. Additionally, the second‐highest accuracy, 99.05%, was obtained from 23, 24, and 27 Relief features. Using 25 features instead of the 1074 features obtained from the proposed CNN model, the classification performance was increased, and the classification duration was reduced. The confusion matrix and ROC curve of 25 Relief features of the proposed COVIDetNet model are given in Figure 8.
FIGURE 8

Confusion matrix and receiver operating characteristic curves for 25 Relief features

Confusion matrix and receiver operating characteristic curves for 25 Relief features As seen in Figure 8, the normal, COVID‐19, and pneumonia samples were identified with 97.5%, 99.7%, and 100% accuracy, respectively. An average score of 99.24% was achieved across the three classes. In addition, with fewer features, such as 25, the classification accuracy was increased by 0.34%. The results in Table 3 and Figure 8 show that the accuracy score is improved with fewer features. In order to more effectively demonstrate the performance of the proposed method, the Conv1, Conv2, Conv3, Conv4, Conv5, fc7, and fc8 layers of the AlexNet model were combined by taking 96, 256, 384, 384, 256, 4096, and 4096 features, respectively, in feature extraction from popular CNN structures. Then, efficient features were selected among 9568 features combined with the embossing algorithm. Finally, the performance of each selected feature vector was calculated with SVM. The performance results of the features selected by the Relief algorithm are given in Table 4. Performance results of effective features with prelearned AlexNet deep features Relief algorithm (%) In Table 4, the accuracy, sensitivity, specificity, precision, and F scores of the features chosen by the Relief algorithm are detailed. The Relief algorithm's maximum accuracy score was 98.66%, attained with 2500–7000 Relief features. While the results in Tables 3 and 4 are compared, the highest accuracy is obtained with 25 features in the proposed method, 2500 features are needed in the AlexNet model. In addition to the classification performances, it is seen in Table 5 that the proposed method is faster than the AlexNet model in the feature extraction and effective features selection process.
TABLE 5

Comparison of the feature extraction and selection times of the proposed and AlexNet models

MethodFeature extracted duration (s)Feature selection duration (s)Total duration (s)Acc (%)
AlexNet + Relief + SVM1421247166898.66
Proposed (COVIDetNet)2592828799.24
Comparison of the feature extraction and selection times of the proposed and AlexNet models As seen in Table 5, the proposed CoviDetNet model can detect COVID‐19 cases in a total of 287 s (approximately 3 min), 259 s for feature extraction and 28 s for selecting effective features. The same operations can be performed for a total of 1668 s (approximately 28 min) with the widely used AlexNet model. In addition, the proposed COVIDetNet method has a 0.58% higher accuracy rate than the AlexNet model.

DISCUSSION

Controlling the rapidly spreading COVID‐19 cases can only be possible by rapidly detecting infected people. However, with the PT‐PCR tests used for case detection, it results in a minimum of 5 h. For this reason, deep learning‐based methods have been developed to obtain fast and highly accurate results under pandemic conditions. Some state‐of‐the‐art methods compared the proposed CoviDetNet model performance results shown in Table 6. Two and three‐class studies are included in the comparisons. The COVIDetNet model proposed in Table 4 and Figure 6 achieved a higher accuracy score than the latest model methods. Using DarkCovidNet for COVID‐19 detection using CXR, Ozturk et al. achieved 98.08% accuracy in binary classification. However, they reached 87.02% in three classifications. In the three‐class studies conducted in 2020, Turkoglu 99.18%, Ucar et al. 98.25%, Wang et al. 93.3%, and Nour et al. achieved an accuracy score of 98.97%. Performance results of state‐of‐the‐art approaches (%) COVID‐19: 125 No findings: 500 COVID‐19: 125 Pneumonia: 500 No findings: 500 COVID‐19: 219 Pneumonia: 4290 Normal: 1583 COVID‐19: 1536 Pneumonia: 1536 Normal: 1536 COVID‐19: 13870 Pneumonia: 5538 Normal: 8066 COVID‐19: 219 Pneumonia: 1345 Normal: 1345 COVID‐19: 500 Pneumonia: 580 Normal: 1541 COVID‐19: 500 Normal: 800 COVID‐19: 500 Pneumonia: 800 Normal: 800 COVID‐19: 1525 Pneumonia: 1525 No Findings:1525 COVID‐19: 1525 Pneumonia: 1525 No Findings:1525 COVID‐19: 900 Pneumonia: 900 Normal: 900 COVID‐19: 125 Pneumonia: 500 Control: 500 COVID‐19: 3616 Pneumonia: 1345 Control: 4000 COVID‐19: 580 Pneumonia: 500 Normal: 1541 Note: Bold values shows the performance results where the proposed method is superior to the existing methods. In studies conducted in 2021, Hussain et al. achieved 99.10% accuracy in binary classification but reached a 94.20% accuracy rate in three classifications in a similar dataset. Cengil and Çınar detected COVID‐19 by combining the features obtained based on transfer learning in AlexNet, Xception, NASNETLarge, and EfficientNet‐B0 models. The AlexNet + NASNetLarge and NASNetLarge + Xception approaches achieved 96.0% and 97.60% accuracy rates. In addition, these two methods reached this accuracy score in 3791 and 4485 s, respectively. Barua et al. used a hybrid model called COVID‐19FclNet9. The fully connected layer of AlexNet, VGG16, and VGG19 CNN models were used to extract deep features using a combined structure. Effective features received in the iterative feature selection algorithm were classified by SVM. Experimental studies used two different datasets, and they achieved an accuracy of 89.6% and 98.84%, respectively. Our study achieved a higher classification performance than existing studies, as shown in Table 6. This confirms that our proposed model is correct and robust. As seen in Table 6, there is no standard data set for detecting COVID‐19 cases. Researchers combined different datasets for studies. For this reason, it may not be correct to say that the studies carried out are utterly superior to each other. Considering this situation, while comparing the methods in Table 6, the number of classes and dataset sizes are also given.

CONCLUSION

COVID‐19 seriously affects public health, the global economy, and social life worldwide. Despite the ongoing vaccination programs for the past year, the number of people affected by COVID‐19 has increased significantly. Due to this increase in cases, the death toll also increased. Therefore, more effective diagnostic methods are needed to assist clinical studies on COVID‐19. For this purpose, a fast and robust model based on the light CNN approach is proposed to detect COVID‐19 cases from CXR images. A scratch‐trained architecture is used in the proposed CNN instead of pretrained models such as ResNet and AlexNet. Deep features were automatically extracted from fully connected and convolution layers in the proposed CNN architecture. The most efficient features among the deep features were selected and classified by SVM with the Relief approach. For the performance analysis of the model, 2621 CXR images obtained from a publicly available dataset were used. The experimental results of the proposed COVIDetNet model are shown below: The proposed CNN model is three times faster than the best AlexNet model in feature extraction. Classification accuracy rates were increased by 0.43% and 0.77%, respectively, with the SVM and SVM + Relief algorithms. In all evaluation criteria, 99.24% accuracy, 99.60% specificity, 99.39% sensitivity, 99.04% acuity, and 99.21% F1 score were obtained. Studies on COVID‐19 case diagnoses have generally focused on accuracy scores. However, it is also crucial to know how long this diagnosis will take in addition to the accuracy score. The proposed CoviDetNet model can diagnose COVID‐19 cases in 287 s. The proposed model can significantly contribute to radiologists and field experts regarding low‐capacity hardware and faster decision‐making during intense COVID‐19 cases. In future works, more refined models will be constituted with big datasets containing many CXR images.

CONFLICT OF INTEREST

The authors declare no conflicts of interest.
  30 in total

1.  COVIDGR Dataset and COVID-SDNet Methodology for Predicting COVID-19 Based on Chest X-Ray Images.

Authors:  S Tabik; A Gomez-Rios; J L Martin-Rodriguez; I Sevillano-Garcia; M Rey-Area; D Charte; E Guirado; J L Suarez; J Luengo; M A Valero-Gonzalez; P Garcia-Villanova; E Olmedo-Sanchez; F Herrera
Journal:  IEEE J Biomed Health Inform       Date:  2020-12-04       Impact factor: 5.772

2.  Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks.

Authors:  Ali Narin; Ceren Kaya; Ziynet Pamuk
Journal:  Pattern Anal Appl       Date:  2021-05-09       Impact factor: 2.580

3.  Automated detection of COVID-19 cases using deep neural networks with X-ray images.

Authors:  Tulin Ozturk; Muhammed Talo; Eylul Azra Yildirim; Ulas Baran Baloglu; Ozal Yildirim; U Rajendra Acharya
Journal:  Comput Biol Med       Date:  2020-04-28       Impact factor: 4.589

4.  COVIDiagnosis-Net: Deep Bayes-SqueezeNet based diagnosis of the coronavirus disease 2019 (COVID-19) from X-ray images.

Authors:  Ferhat Ucar; Deniz Korkmaz
Journal:  Med Hypotheses       Date:  2020-04-23       Impact factor: 1.538

5.  E-DiCoNet: Extreme learning machine based classifier for diagnosis of COVID-19 using deep convolutional network.

Authors:  R Murugan; Tripti Goel
Journal:  J Ambient Intell Humaniz Comput       Date:  2021-01-02

6.  CoviDetNet: A new COVID-19 diagnostic system based on deep features of chest x-ray.

Authors:  Muzaffer Aslan
Journal:  Int J Imaging Syst Technol       Date:  2022-06-10       Impact factor: 2.177

7.  Covert COVID-19 and false-positive dengue serology in Singapore.

Authors:  Gabriel Yan; Chun Kiat Lee; Lawrence T M Lam; Benedict Yan; Ying Xian Chua; Anita Y N Lim; Kee Fong Phang; Guan Sen Kew; Hazel Teng; Chin Hong Ngai; Li Lin; Rui Min Foo; Surinder Pada; Lee Ching Ng; Paul Anantharajah Tambyah
Journal:  Lancet Infect Dis       Date:  2020-03-04       Impact factor: 25.071

8.  Covid-19: automatic detection from X-ray images utilizing transfer learning with convolutional neural networks.

Authors:  Ioannis D Apostolopoulos; Tzani A Mpesiana
Journal:  Phys Eng Sci Med       Date:  2020-04-03

9.  Real-time forecasts of the COVID-19 epidemic in China from February 5th to February 24th, 2020.

Authors:  K Roosa; Y Lee; R Luo; A Kirpich; R Rothenberg; J M Hyman; P Yan; G Chowell
Journal:  Infect Dis Model       Date:  2020-02-14
View more
  1 in total

1.  CoviDetNet: A new COVID-19 diagnostic system based on deep features of chest x-ray.

Authors:  Muzaffer Aslan
Journal:  Int J Imaging Syst Technol       Date:  2022-06-10       Impact factor: 2.177

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.