Mesut Toğaçar1, Burhan Ergen2, Zafer Cömert3. 1. Department of Computer Technology, Vocational School of Technical Sciences, Fırat University Elazig, Turkey. Electronic address: mtogacar@firat.edu.tr. 2. Department of Computer Engineering, Faculty of Engineering, Fırat University Elazig, Turkey. Electronic address: bergen@firat.edu.tr. 3. Department of Software Engineering, Faculty of Engineering, Samsun UniversitySamsun, Turkey. Electronic address: zcomert@samsun.edu.tr.
Abstract
Coronavirus causes a wide variety of respiratory infections and it is an RNA-type virus that can infect both humans and animal species. It often causes pneumonia in humans. Artificial intelligence models have been helpful for successful analyses in the biomedical field. In this study, Coronavirus was detected using a deep learning model, which is a sub-branch of artificial intelligence. Our dataset consists of three classes namely: coronavirus, pneumonia, and normal X-ray imagery. In this study, the data classes were restructured using the Fuzzy Color technique as a preprocessing step and the images that were structured with the original images were stacked. In the next step, the stacked dataset was trained with deep learning models (MobileNetV2, SqueezeNet) and the feature sets obtained by the models were processed using the Social Mimic optimization method. Thereafter, efficient features were combined and classified using Support Vector Machines (SVM). The overall classification rate obtained with the proposed approach was 99.27%. With the proposed approach in this study, it is evident that the model can efficiently contribute to the detection of COVID-19 disease.
Coronavirus causes a wide variety of respiratory infections and it is an RNA-type virus that can infect both humans and animal species. It often causes pneumonia in humans. Artificial intelligence models have been helpful for successful analyses in the biomedical field. In this study, Coronavirus was detected using a deep learning model, which is a sub-branch of artificial intelligence. Our dataset consists of three classes namely: coronavirus, pneumonia, and normal X-ray imagery. In this study, the data classes were restructured using the Fuzzy Color technique as a preprocessing step and the images that were structured with the original images were stacked. In the next step, the stacked dataset was trained with deep learning models (MobileNetV2, SqueezeNet) and the feature sets obtained by the models were processed using the Social Mimic optimization method. Thereafter, efficient features were combined and classified using Support Vector Machines (SVM). The overall classification rate obtained with the proposed approach was 99.27%. With the proposed approach in this study, it is evident that the model can efficiently contribute to the detection of COVID-19 disease.
The new Coronavirus (COVID-19) is an acute deadly disease that originated from Wuhan province, China in December 2019 and spread globally. COVID-19 outbreak has been of great concern to the health community because no effective cure has been discovered [1]. The biological structure of COVID-19 comprises of a positive-oriented single-stranded RNA-type, and it is difficult to treat the disease owing to its mutating feature. Medical professionals globally are undergoing intensive research to develop an effective cure for the disease. Presently, COVID-19 is the primary cause of thousands of deaths globally, and major deaths are in the USA, Spain, Italy, China, the UK, Iran, etc. Many types of coronavirus exist, and these viruses are commonly seen in animals. COVID-19 has been discovered in human, bat, pig, cat, dog, rodent, and poultry. Symptoms of COVID-19 include sore throat, headache, fever, runny nose, and cough. The virus can provoke the death of people with weakened immune systems [2,3]. COVID-19 is transmitted from person to person mostly by physical contact. Generally, healthy people can be infected through breath contact, hand contact, or mucous contact with people carrying COVID-19 [4].Recently, artificial intelligence (AI) has been widely used for the acceleration of biomedical research. Using deep learning approaches, AI has been used in many applications such as image detection, data classification, image segmentation [5,6]. Peopleinfected by COVID-19 may suffer from pneumonia because the virus spreads to the lungs. Many deep learning studies have detected the disease using a chest X-ray image data approach [7]. A previous study has classified the pneumonia X-ray images using three different deep learning models [8] namely the fine-tuned model, model without fine-tuning, and the model trained from scratch. By using the ResNet model, they classified the dataset into multiple labels such as age, gender, etc. They also used the Multi-Layer Perceptron (MLP) as a classification method and achieved an average of 82.2% accuracy. Samir Yadav et al. [9] performed a classification algorithm using pneumonia data, SVM as a classification method, and InceptionV3, VGG-16 models as a deep learning approach. In their study, the dataset is divided into three classes: normal, bacterial pneumonia, and viral pneumonia to improve the contrast and brightness zoom settings with the augmentation method for each image in the dataset. The best classification achievement was 96.6%. Rahib Abiyev et al. [10] used the Backpropagation Neural Network and Competitive Neural Network models to classify pneumonia data. Using pneumonia and normal chest X-ray images, they set 30% of the dataset as test data and compared the proposed approach with the existing CNNs. They achieved 89.57% classification success. Okeke Stephen et al. [11] proposed a deep learning model to classify the pneumonia data from scratch to train the data. Their proposal consists of convolution layers, dense blocks, and flatten layers. The input size of the model was 200 × 200 pixels to determine the possibilities of classification using the sigmoid function. The success rate was 93.73% in pneumonia from X-ray images. Vikash Chouhan et al. [12] detected the images of pneumonia using deep learning models, three classes of the dataset: normal, virus pneumonia, and bacterial pneumonia images. In the first instance, they carried out a set of preprocessing procedures to remove noise from the images. Then, they applied the augmentation technique to each image and used a transfer learning to train the models. The overall classification accuracy was 96.39%.In this study, we used COVID-19 chest images dataset, pneumonia chest images, and normal chest images. We preprocessed each image before being trained with deep learning models. In this preprocessing, the dataset was reconstructed using the Fuzzy technique and Stacking technique. Then, we trained the three datasets using the MobileNetV2 and SqueezeNet deep learning models and classified the models by the SVM method. The remainder of this study is structured as follows: In Section 2, we discuss the structure of the dataset, technique, method, deep learning models, and optimization algorithm. Experimental analysis is mentioned in Section 3. Section 4, 5 present the discussion and conclusion, respectively.
Dataset, models, methods, and techniques
Dataset
In the experimental analysis, we use the three classes of datasets that are accessible publicly. These classes are normal, pneumonia, and COVID-19 chest images. All datasets are X-ray images, and each image is converted to JPG format. Since COVID-19 is a new disease, the number of images related to this virus is limited. For this study, we combined two publicly accessible databases consisting of COVID-19 images. The first COVID-19 dataset was shared on the GitHub website by a researcher named Joseph Paul Cohen from the University of Montreal. After the experts checked the images, they were made available to the public. In the Joseph Paul Cohen dataset, image types are MERS, SARS, COVID-19, etc. The data of 76 images labeled with COVID-19 were selected for this study [13]. The second COVID-19 dataset consists of the images created by a team of researchers from Qatar University, medical doctors from Bangladesh, and collaborators from Pakistan and Malaysia. The second COVID-19 dataset is available on the Kaggle website, and the current version has 219 X-ray images [14]. For this study, two datasets containing COVID-19 images were combined, and a new dataset consisting of 295 images was created. COVID-19 virus causes pneumonia in the affected individuals and can provoke death if the lungs are permanently damaged [15]. The second dataset is important in this study to compare COVID-19 chest images using deep learning models. The second dataset consists of normal chest images and pneumonia chest images. Pneumonia chest images include both virus and bacteria types, and these images are taken from 53 patients. The images were created by experts and shared publicly [16].The combined dataset consists of three classes. Information about the classes of the dataset and the number of images in the classes are as follows: We collect a total of 295 images in COVID-19 class. Normal class X-ray images are 65 in total, and pneumonia class X-ray images are 98. The total number of images of the dataset is 458. In the experimental analysis, 70% of the dataset was used as training data, and 30% was used as test data. In the last step of the experiment, the k-fold cross-validation method was used for stacked images. Sample images of the dataset are shown in Fig. 1
.
Fig. 1
The sample images used in the experimental analysis of this study; (a) COVID-19 chest images, (b) normal chest images, (c) pneumonia chest images.
The sample images used in the experimental analysis of this study; (a) COVID-19 chest images, (b) normal chest images, (c) pneumonia chest images.
Deep learning model: MobileNetV2
MobileNet is a deep learning model intended to be used in low hardware cost devices. Object identification, segmentation, and classification can be performed using the MobileNet model. The MobileNet model is known as MobileNetV1, and the MobileNetV2 model is developed from the MobileNetV1. Compared to the MobileNetV2 model with the previous version, this new model offers the biggest contribution to the problems of linearity between layers. If linear bottlenecks occur between the layers, the problems are fixed in this version with shortcuts [17]. Fig. 2
shows the architecture design of the MobileNetV2 model. Its input size is 224 × 224 pixels, and its architecture comprises of in-depth (DW) separable filters and combination of steps. The model performance increases as it examines the DW, and the input features are divided into two layers. Each layer is subdivided into the next layer by combining it with the output features until the process is completed. MobileNetV2 model uses the ReLU function between layers [18]. Thus, it enables the nonlinear outputs from the previous layer to be linearized and transferred as input to the next layer. The model continues its training process until a comfortable step is reached. In this model, the convolutional layers circulate filters over input images and create activation maps. The activation maps contain the features extracted from the input images, and these features are transferred to the next layer. Pooling layers are also used in the MobileNetV2 model. The matrices obtained through these layers are converted into smaller dimensions [19].
Fig. 2
The general design of the MobileNetV2 model used in this study [20].
The general design of the MobileNetV2 model used in this study [20].MobileNetV2 model was used as pre-trained, and the SVM method was used in the classification phase. Besides, other important parameters of the MobileNetV2 model are given in Table 1
and Table 2
. All default parameter values were used for the MobileNetV2 model without any change.
Table 1
General structure and parameters of MobileNetV2 architecture.
Type
Stride
Filter Size
Input Size
Convolution
2 × 2
3 × 3 × 3 × 32
224 × 224 × 3
Convolution DW
1 × 1
3 × 3 × 32
112 × 112 × 32
Convolution
1 × 1
1 × 1 × 32 × 64
112 × 112 × 32
Convolution DW
2 × 2
3 × 3 × 64
112 × 112 × 64
Convolution
1 × 1
1 × 1 × 64 × 128
56 × 56 × 64
Convolution DW
1 × 1
3 × 3 × 128
56 × 56 × 128
Convolution
1 × 1
1 × 1 × 128 × 128
56 × 56 × 128
Convolution DW
2 × 2
3 × 3 × 128
56 × 56 × 128
Convolution
1 × 1
1 × 1 × 128 × 256
28 × 28 × 128
Convolution DW
1 × 1
3 × 3 × 256
28 × 28 × 256
Convolution
1 × 1
1 × 1 × 256 × 256
28 × 28 × 256
Convolution DW
2 × 2
3 × 3 × 256
28 × 28 × 256
Convolution
1 × 1
1 × 1 × 256 × 512
14 × 14 × 256
5 × Convolution DW
1 × 1
3 × 3 × 512
14 × 14 × 512
5 × Convolution
1 × 1
1 × 1 × 512 × 512
14 × 14 × 512
Table 2
Other parameters of the MobileNetV2 model and preferred values in this study.
Software Used
Model
Image Size
Optimization
Momentum
Decay
Beta
MiniBatch
LearningRate
MATLAB
MobileNetV2
224 × 224
Stochastic Gradient Descent (SGD)
0.9
1e-6
–
64
10–5
General structure and parameters of MobileNetV2 architecture.Other parameters of the MobileNetV2 model and preferred values in this study.
Deep learning model: SqueezeNet
SqueezeNet, an in-depth learning model of input size of 224 × 224 pixels, comprises of convolutional layers, pooling layers, ReLU, and Fire layers. The SqueezeNet does not have fully connected layers and dense layers. However, Fire layers perform the functions of these similar layers. The major benefit of this model is that it performs analyses successfully by reducing the number of parameters, thereby decreasing the model size capacity. SqueezeNet model produced more successful results, approximately 50 times fewer parameters than the AlexNet model, thereby reducing the cost of the model [21]. Fig. 3
presents the model design.
Fig. 3
The general design of the SqueezeNet model used in this study [21].
The general design of the SqueezeNet model used in this study [21].While the information about the layers is presented in the MobileNetV2) model, the Fire (F2, F3, ..., F9) layers that look like a new layer consisting of two parts namely the Compression and Expansion parts. This model uses only a 1 × 1 convolutional filter to the input image in the Compression part of the Fire layer. In the Expansion part, it uses both 1 × 1 and 3 × 3 convolutional filters to the input image. The Compression part and the Expansion part keep the same feature map size. In the Compression part, the depth of the input image is reduced and then increased (bottleneck). In the Expansion part, the depth is increased [21,22]. Table 3
presents the layers and default parameter values of the model, these values are used without changes. Other important parameters of the SqueezeNet model are given in Table 4
.
Table 3
General structure and parameters of SqueezeNet architecture.
Type
Stride
Filter Size
Output Size
Input
–
–
224 × 224 × 3
Convolution
2
96 × 96 × 7
109 × 109 × 96
Pooling
2
3 × 3
54 × 54 × 96
Fire 2
–
16 × 16 × 1, 64 × 1 × 1, 64 × 3 × 3
54 × 54 × 128
Fire 3
–
16 × 16 × 1, 64 × 1 × 1, 64 × 3 × 3
54 × 54 × 128
Fire 4
–
32 × 1 × 1, 128 × 1 × 1, 128 × 3 × 3
54 × 54 × 256
Pooling
2
3 × 3
27 × 27 × 256
Fire 5
–
32 × 1 × 1, 128 × 1 × 1, 128 × 3 × 3
27 × 27 × 256
Fire 6
–
48 × 1 × 1, 192 × 1 × 1, 192 × 3 × 3
27 × 27 × 384
Fire 7
–
48 × 1 × 1, 192 × 1 × 1, 192 × 3 × 3
27 × 27 × 384
Fire 8
–
64 × 1 × 1, 256 × 1 × 1, 256 × 3 × 3
27 × 27 × 512
Pooling
2
3 × 3
13 × 13 × 128
Fire 9
–
64 × 1 × 1, 256 × 1 × 1, 256 × 3 × 3
13 × 13 × 512
Convolution
1
6 × 13 × 13
13 × 13 × 6
Pooling
–
13 × 13
1 × 1 × 6
Table 4
Other parameters of the SqueezeNet model and preferred values in this study.
Software Used
Model
Image Size
Optimization
Momentum
Decay
Beta
MiniBatch
LearningRate
MATLAB
SqueezeNet
224 × 224
SGD
0.9
1e-6
–
64
10–5
General structure and parameters of SqueezeNet architecture.Other parameters of the SqueezeNet model and preferred values in this study.
Classification method: SVM and optimization method: SGD
SVM is a machine learning used for regression and classification analysis. This method uses a hyper-plane line in the classification process to separate the features in the data classes. It chooses a location away from the features of the classes to determine a line. The distances of each class are measured according to the hyper-plane determined by the SVM method. The one with the highest voting score is transferred to the class labeled as the highest voting score. Fig. 4
shows the design of the SVM method for the classification process. The functions of the mathematical operations are given in Eq. (1), Eq. (2) and Eq. (3). Here, and represent the coordinate points of the features in the hyperplane. parameter represents margin width, and parameter represents bias value [23,24].
Fig. 4
Design of SVM method for multiple classification process.
Design of SVM method for multiple classification process.Deep learning models use the optimization methods, and these facilitate learning trends of models. Stochastic Gradient Descent (SGD) optimization is the method that updates the weight parameters in the model structure at every iteration. The models provide better training during each iteration. However, the SGD does not use all the images input into the model while updating the parameters. It performs this operation using only the images it randomly determines. This lowers the cost of the model, and it offers a faster training process. Eq. (4) shows the mathematical function that performs the weight parameter updates in the SGD. Where represents the weight parameter. This represents the coordinates of the features extracted in the and input images and represents α learning rate [25].In this study, the SVM method was preferred because:It has a strong potential to provide solutions to the data analysis problems encountered in daily life,It is widely used for the remote pattern recognition and classification problems to successfully execute multiple classification processes [26,27], andIt gives the best classification performance among other machine learning methods (discriminant analysis, nearest neighbor, etc.).Moreover, the Linear SVM was preferred owing to its best performance such as cubic, linear, quadratic, etc. Preferred parameter values in the Linear SVM method were the kernel scale that was parameter automatically selected. The box constraint level parameter value was chosen, and the multiclass method parameter of one-vs-one was selected.
Reconstructing images: fuzzy color technique
The Fuzzy concept is accepted based on its degree of accuracy, and its next degree is uncertain. Fuzzy Color algorithms play an important role in image analysis, and the obtained results depend on the similarity/difference functions used for color separation. In the fuzzy color technique, each of the input images contains three input variables (red, green, and blue - RGB). As a result of this process, a single output variable is passed. The input and output values are determined according to the training data [28,29].The logic behind the Fuzzy Color technique is to separate the input data into blurred windows. Each pixel in the image has a membership degree to each window, and membership degrees are calculated based on the distance between the window and the pixel. Image variance is obtained with membership degrees. The Fuzzy Color technique is to create the finishing output-input. In this step, the weights of the images of each blurred window are summed, and the output image is created from the average. Here, the weight value of each pixel is expressed as the degree of membership [28,30]. We recreated the original dataset using the Python codes with the Fuzzy Color technique [31]. Fig. 5
shows the structure of the data image.
Fig. 5
Sub-data samples of the original dataset obtained by the Fuzzy Color technique.
Sub-data samples of the original dataset obtained by the Fuzzy Color technique.
Reconstructing images: stacking technique
Image stacking is a digital image processing technique that combines multiple images shot or is reconstructed at different focal distances. This is a technique used to improve the quality of the images in the dataset. This technique aims to eliminate the noises from the original image by combining at least two images in a row and dividing the image into two parts. These parts are background and overlay. While the first images are processed in the background, the second is overlaid on the image placed in the background. Here, parameters such as opacity, contrast, brightness, and combining ratio of the two images are important. The more accurately these ratios, the more amount of noise is reduced in the images, and the higher the quality ratio [32].In this study, Python and Pillow library was used for the Stacking Technique [33]. Here, the original dataset was stacked on the reconstructed dataset using the Fuzzy technique. The successful result of the Fuzzy technique will contribute to the success of the stacking technique. The parameter values preferred in the stacking technique were the opacity value, with the value of 0.6, the contrast value was 1.5, the brightness value was set to −80, and the combined ratio was chosen as 50%. These values can be varied for another dataset. We evaluate the various stages for dataset images and determine that these values are the most efficient of the dataset. Hence, we used them in the experimental analysis. Moreover, the original dataset was placed in the background, and the structured dataset was placed in the overlay. A combined representation of the original dataset using the stacking technique and dataset structured with the Fuzzy technique is shown in Fig. 6
.
Fig. 6
Sub-image samples obtained by the Stacking technique.
Sub-image samples obtained by the Stacking technique.
Social Mimic Optimization
Emotions such as morale and happiness can be transfer from person to person. Emotional imitation happens consciously or unconsciously. This situation is related to neurons in the human brain, that is, the nervous system. Some of these neurons are responsible for perceiving facial expressions such as our eyebrows sharpening or laughing, etc. These neurons also act as interbrain communication and generate a response to this effect. In other words, social imitation is the condition of adopting the behavior, conversation, or dressing of another person, which one sees as a guide. Based on this, an advanced is made towards a better understanding of human behavioral activities [34].Social Mimic Optimization (SMO) is a method developed by inspiring people to imitate other people. This is inspired by imitating the behavior of people in society. Locally, each problem produces a solution using the SMO algorithm, which moves towards the global solution. Each solution determines the difference in the global value by comparing the local value obtained in the last iteration. This is applied to the problem parameters to find the solution randomly based on the difference in the value obtained. While the term "" in SMO expresses the population, the “” parameter expresses the best global value, and the “” parameter represents the number of iterations. The number of followers is obtained by multiplying the population with the number of decision variables (). Decision variables have a lower bound () and upper bound () value. Eqs. (5), (6), (7) are used to implement the SMO algorithm [34]. For this study, the population size of 20 was selected, the maximum iteration parameter was selected as 10. The best global value parameter was taken as 1000 for the start, and the parameter value of 10 was chosen.
Proposed approach
The proposed approach is designed to perform the classification of chest images based on the dataset types. This is an approach to separate chest images of COVID-19virus infection from normal breast images and pneumonia. The original dataset is passed through the front image processing steps. In the first step, the original dataset was recreated with the Fuzzy Color technique, aimed to remove the noise in the original images. In the second step, using the original dataset Stacking technique, each fuzzy color image was combined with original images, and a new dataset was created. The aim was to create a better data quality image. The two deep learning models were used, and the stacked dataset was trained with MobileNetV2 and SqueezeNet deep learning models. Using the SMO algorithm, the 1000-features obtained by the models were used to extract efficient features. By combining the efficient features, the SVM method that produced successful results in multiple classifications was used as the classification process. Fig. 7
shows the overall design of the proposed approach.
Training and test success graphs of the SqueezeNet model; (a) original dataset, (b) dataset restructured using the Fuzzy technique, (c) dataset combined using the Stacking technique.
Fig. 9
Confusion matrices of the SqueezeNet model; (a) original dataset, (b) dataset restructured using the Fuzzy technique, (c) dataset combined using the Stacking technique.
Table 5
Metric values of the confusion matrix obtained by the SqueezeNet model.
Model&Dataset Type
Classes
F-Scr. (%)
Se. (%)
Sp. (%)
Pre. (%)
Acc. (%)
OverallAcc. (%)
SqueezeNet&Original Data
COVID-19
96.59
96.59
90.91
96.59
95.04
84.56
Normal
44.44
31.57
98.19
75
88.46
Pneumonia
69.56
82.75
85.04
60
84.56
SqueezeNet&Structured dataset (Fuzzy Technique)
COVID-19
99.43
98.86
100
100
99.24
95.58
Normal
87.80
94.73
96.55
81.82
96.29
Pneumonia
89.29
86.21
98.13
92.59
95.59
SqueezeNet&Stacked dataset (Stacked Technique)
COVID-19
99.44
100
97.78
98.88
99.25
97.06
Normal
91.89
89.47
99.14
94.44
97.78
Pneumonia
93.10
93.10
98.13
93.10
97.06
Fig. 10
Training and test success graphs of the MobileNetV2 model; (a) original dataset, (b) dataset restructured using the Fuzzy technique, (c) dataset combined using the Stacking technique.
Fig. 11
Confusion matrices of the MobileNetV2 model; (a) original dataset, (b) dataset restructured using the Fuzzy technique, (c) dataset combined using the Stacking technique.
Table 6
Metric values of the confusion matrix obtained by the MobileNetV2 model.
Model&Dataset Type
Classes
F-Scr. (%)
Se. (%)
Sp. (%)
Pre. (%)
Acc. (%)
OverallAcc. (%)
MobileNetV2&Original Data
COVID-19
99.43
100
97.72
98.87
99.24
96.32
Normal
89.47
89.47
98.27
89.47
97.03
Pneumonia
91.22
89.65
98.13
92.85
96.32
MobileNetV2&Structured dataset (Fuzzy Technique)
COVID-19
98.29
97.73
97.87
98.85
97.78
97.05
Normal
94.44
89.47
100
100
98.51
Pneumonia
95.08
100
97.17
90.63
97.78
MobileNetV2&Stacked dataset (Stacked Technique)
COVID-19
98.31
98.86
95.74
97.75
97.78
97.06
Normal
94.74
94.74
99.13
94.74
98.51
Pneumonia
94.74
93.10
99.06
96.43
97.78
Training and test success graphs of the SqueezeNet model; (a) original dataset, (b) dataset restructured using the Fuzzy technique, (c) dataset combined using the Stacking technique.Confusion matrices of the SqueezeNet model; (a) original dataset, (b) dataset restructured using the Fuzzy technique, (c) dataset combined using the Stacking technique.Metric values of the confusion matrix obtained by the SqueezeNet model.Training and test success graphs of the MobileNetV2 model; (a) original dataset, (b) dataset restructured using the Fuzzy technique, (c) dataset combined using the Stacking technique.Confusion matrices of the MobileNetV2 model; (a) original dataset, (b) dataset restructured using the Fuzzy technique, (c) dataset combined using the Stacking technique.Metric values of the confusion matrix obtained by the MobileNetV2 model.In the second step of this experiment, the k-fold cross-validation method was used for the stacked dataset, classified by using the SVM method. In the first step of the experiment, 30% of the dataset was used as test data. To confirm the validity of the results of the first step, the dataset was separated using a k-fold cross-validation method. For two CNN models, the k-fold value was adjusted to five. In the second step, the overall accuracy rate achieved with the SqueezeNet model was 95.85%. In the first step of this model, the classification rate with the stacked dataset (with 30% test data) was 97.06%. In the analyzes performed in the second step, the SqueezeNet model produced stable results in both steps. In the second step, the overall accuracy rate was 96.28%, obtained from the MobileNetV2 model trained with the stacked dataset. In the first step of the MobileNetV2 model, the classification rate was 97.06%, achieved with the stacked dataset (with 30% test data). Fig. 12
shows the confusion matrices of the analysis performed in the second step, and Table 7
gives the metric values obtained from the confusion matrices. As a result, all analyzes performed in the second step produce a stable result compared to the results obtained in the first step. The results obtained by the 5-fold cross-validation confirmed the reliability of the proposed approach compared to the results obtained from the previous step.
Fig. 12
Confusion matrices with the method of 5-fold cross-validation of stacked data; (a) using the SqueezeNet model, (b) using the MobileNetV2 model.
Table 7
Analysis results of the stacked dataset with the 5-fold cross-validation method.
Model&Dataset Type
Classes
F-Scr. (%)
Se. (%)
Sp. (%)
Pre. (%)
Acc. (%)
OverallAcc. (%)
SqueezeNet&Stacked dataset (Stacked Technique)
COVID-19
98.82
98.98
97.35
98.65
98.43
95.85
Normal
88
84.62
98.71
96.70
96.70
Pneumonia
92
93.88
97.20
90.19
96.48
MobileNetV2&Stacked dataset (Stacked Technique)
COVID-19
98.81
98.31
98.69
99.32
98.44
96.28
Normal
90.91
92.31
98.20
89.55
97.35
Pneumonia
92.39
92.86
97.77
91.92
96.71
Confusion matrices with the method of 5-fold cross-validation of stacked data; (a) using the SqueezeNet model, (b) using the MobileNetV2 model.Analysis results of the stacked dataset with the 5-fold cross-validation method.In the third step, the SMO algorithm was applied to the feature sets obtained from the stacked dataset trained by CNN models. The dataset contains 1000 features, extracted by two CNN models. The dataset was obtained as a file with a '*.mat’ extension from MATLAB. These feature sets were obtained from the Pool10 layer in the SqueezeNet model and the Logits layer in the MobileNetV2 model. Choosing efficient features with the SMO algorithm offers a total of 800 column numbers as we have set the maximum features selection to 800, and some of these column numbers are repeated using the SMO algorithm. The feature column numbers selected by CNN models are less than 800. Using an efficient optimization method with a few possible features, this step aimed to contribute to the classification success. The third step consists of four stages. In the first stage, the SMO algorithm was applied to the 1000-feature set, obtained by training the stacked dataset using the SqueezeNet model, and column numbers with efficient features were selected. The SqueezeNet model selected 694 efficient features from 1000 features, and the selected features were classified by the SVM method, which achieved an overall 97.81% accuracy rate. Also, the classification success in the detection of COVID-19 data with the SqueezeNet model was 100%. In the second stage, the SMO algorithm was applied to the 1000-feature set obtained from the stacked dataset with the MobileNetV2 model, and column numbers with efficient features were selected. The MobileNetV2 model selected 663 efficient features from 1000 features, and the efficient features were classified by the SVM method to achieve an overall 98.54% accuracy rate. Moreover, the classification success was 99.26% for the detection of COVID-19 data with the MobileNetV2 model. In the third stage, 694 efficient features obtained by the SMO algorithm using the SqueezeNet model were combined with the 663 efficient features obtained by the SMO algorithm and the MobileNetV2 model. The aim was to determine whether combining efficient features contributes to better classification performance. By combining two efficient features, a new dataset with 1357 features was obtained. Classified using the SVM method, 30% of the combined feature set was set as test data. The overall accuracy rate obtained from this classification was 99.27%. This result showed that combining selected features with the SMO algorithm contributes to the classification success. In the third stage, the accuracy rate in the classification of COVID-19 data was 100%, and the success rate was 99.27% in the classification of Normal chest images and Pneumonia chest images. In the fourth stage, the cross-validation method (k-fold = 5) was applied to the combined feature set (dataset with 1357 features). This stage aimed to verify the validity of the success achieved in the third stage using the cross-validation method. The overall classification accuracy rate was 98.25% by using the SVM method. The result obtained in the last stage showed that the proposed approach is reliable and valid. In the fourth stage, the success rate was 99.34% in the classification of COVID-19 data, and the success rate was 98.68% in the classification of Normal chest images. The classification success rate was 98.47% in Pneumonia chest images. Fig. 13
shows the confusion matrices of the analysis performed in the third step of the experiment, and Table 8
gives the values of the metric parameters.
Fig. 13
Confusion matrices obtained using the SMO method; (a) with the SqueezeNet model, (b) with the MobileNetV2 model, (c) with combining the features from the SqueezeNet model and the MobileNetV2 model (30% test data), (d) with combining the features from the SqueezeNet model and the MobileNetV2 model (k fold value = 5).
Table 8
Metric values obtained using the SMO method.
Model/Dataset Type
Classes
Total of Features
Test Data (%)
F-Scr. (%)
Se. (%)
Sp. (%)
Pre. (%)
Acc. (%)
OverallAcc. (%)
SqueezeNet
COVID-19
694
30
100
100
100
100
100
97.81
Normal
92.31
90
99.14
94.74
97.81
Pneumonia
94.91
96.55
98.15
93.33
97.81
MobileNetV2
COVID-19
663
30
99.44
100
97.87
98.89
99.26
98.54
Normal
94.44
89.47
100
100
98.54
Pneumonia
98.31
100
99.06
96.67
99.26
SqueezeNet&MobileNetV2 (Combined Features Set)
COVID-19
1357
30
100
100
100
100
100
99.27
Normal
97.43
95
100
100
99.27
Pneumonia
98.30
100
99.07
96.67
99.27
COVID-19
1357
k-fold (k = 5)
99.49
99.32
99.37
99.66
99.34
98.25
Normal
95.31
93.85
99.49
96.82
98.68
Pneumonia
96.48
97.96
98.61
95.05
98.47
Confusion matrices obtained using the SMO method; (a) with the SqueezeNet model, (b) with the MobileNetV2 model, (c) with combining the features from the SqueezeNet model and the MobileNetV2 model (30% test data), (d) with combining the features from the SqueezeNet model and the MobileNetV2 model (k fold value = 5).Metric values obtained using the SMO method.In the proposed approach, the contribution of the SMO algorithm was recorded for the improvement of classification performance. The codes and analysis of results of the SMO algorithm are available in the web address of the Open Source Code. Available in the web address is also feature sets obtained in the experiment and related source codes.
Discussion
The number of confirmed COVID-19 cases has exceeded millions worldwide with thousands of confirmed deaths. The World Health Organization has declared that COVID-19 is a global epidemic [37]. Using the proposed approach, we performed the detection of COVID-19 from the X-ray image data. We compared COVID-19 chest data with that of pneumonia and normal chest data since pneumonia is one of the symptoms of COVID-19. The major challenge we encountered is that the publication of COVID-19 images is still limited. Moreover, previous studies on the detection of COVID-19 using deep learning are non-existence. Hence, we fill the gap in the literature.Although the limited dataset is used, this study contributes to the classification of the dataset, using the image preprocessing to determine the data classes. Other techniques can also be used instead of the Fuzzy Color technique. We paid attention to the similarity between the structured image and the original stack image. If we configured the image in a different format (such as resolution status or color pixel status), we would not have achieved the success achieved in this study.The advantages of the proposed approach are as follows:It provides a 100% success rate in detecting the disease by examining the X-ray images of COVID-19patients.The analysis can be carried out using AI, and the proposed approach can be integrated into portable smart devices (mobile phones, etc.)The deep learning models (MobileNetV2 and SqueezeNet) used in the proposed approach have fewer parameters compared to other deep models. This helps to gain speed and time performance. Besides, using the SMO algorithm, CNN models save time and speed during the process.It minimizes the interference in every image in the dataset and provides efficient features with stacking technique.The disadvantages of the proposed approach are as follows:If the sizes of the input images in the dataset are different, a complete success may not be achieved. Irrespective of the resize parameter, it is still a challenge for the proposed approach to deal with very low-resolution images.In the Stacking technique, the resolution dimensions of the original images and the structured images must be the same.We ensure that the number of pneumonia and normal chest images are close to that of COVID-19 chest images. Since we presume that the image classes found in an unbalanced number cannot contribute to the success of the model, we still achieved an overall 99.27% accuracy in the classification process.In the proposed model, the end-to-end learning scheme has been exploited, which is one of the great advantages of CNN models. The pathologic patterns were detected and identified by using the activation maps that kept the discriminative features of the input data. In this manner, the tedious and labor-intensive feature extraction process was isolated; a highly sensitive decision tool was ensured.
Conclusion
Peopleinfected with COVID-19 are likely to suffer permanent damage in the lungs, which can later provoke death. This study aimed to distinguish people with damaged lungs owing to COVID-19 from normal individuals or pneumonia (not infected by COVID-19). The detection of COVID-19 was carried out using deep learning models. Since it is important to detect COVID-19 that spread rapidly and globally, AI techniques are used to perform this accurately and quickly. One of the novelty aspects of the proposed approach is to apply the pre-processing steps to the images. When using pre-processing steps, more efficient features are extracted from the image data. With the stacking technique, each pixel of equivalent images is superimposed, and the pixels with low efficiency is increased. With the proposed approach, efficient features were extracted using the SMO algorithm. The model was aimed to produce faster and more accurate results. Another innovative aspect is that the feature sets obtained with SMO are combined to improve the classification performance. We also demonstrate the usability of our approach to smart mobile devices with the MobileNetV2 model as it can be analyzed on mobile devices without using any hospital devices. The 100% success was achieved in the classification of COVID-19 data, and 99.27% success was achieved in the classification of Normal and Pneumonia images.In future studies, deep learning-based analyzes will be carried out using data images of other organs affected by the virus, in line with the views of COVID-19 specialists. We plan to develop a future approach using different structuring techniques to enhance the datasets. As the data related to the factors influencing the virus in human chemistry (e.g. blood group, RNA sequence, age, gender, etc.) are available, we will produce a solution-oriented analysis using AI.
Open Source Code
Information about Python and MATLAB software source codes, datasets, and related analysis results used in this study are given in this web link. https://github.com/mtogacar/COVID_19.
Funding
There is no funding source for this article.
Ethical approval
This article does not contain any data, or other information from studies or experimentation, with the involvement of human or animal subjects.
Declaration of competing interest
The authors declare that there is no conflict to interest related to this paper.
Authors: Camilla Rothe; Mirjam Schunk; Peter Sothmann; Gisela Bretzel; Guenter Froeschl; Claudia Wallrauch; Thorbjörn Zimmer; Verena Thiel; Christian Janke; Wolfgang Guggemos; Michael Seilmaier; Christian Drosten; Patrick Vollmar; Katrin Zwirglmaier; Sabine Zange; Roman Wölfel; Michael Hoelscher Journal: N Engl J Med Date: 2020-01-30 Impact factor: 91.245
Authors: Mohammed S Alqahtani; Mohamed Abbas; Ali Alqahtani; Mohammad Alshahrani; Abdulhadi Alkulib; Magbool Alelyani; Awad Almarhaby; Abdullah Alsabaani Journal: Diagnostics (Basel) Date: 2021-05-10