Literature DB >> 33041635

A novel comparative study for detection of Covid-19 on CT lung images using texture analysis, machine learning, and deep learning methods.

Huseyin Yasar1, Murat Ceylan2.   

Abstract

The Covid-19 virus outbreak that emerged in China at the end of 2019 caused a huge and devastating effect worldwide. In patients with severe symptoms of the disease, pneumonia develops due to Covid-19 virus. This causes intense involvement and damage in lungs. Although the emergence of the disease occurred a short time ago, many literature studies have been carried out in which these effects of the disease on the lungs were revealed by the help of lung CT imaging. In this study, 1.396 lung CT images in total (386 Covid-19 and 1.010 Non-Covid-19) were subjected to automatic classification. In this study, Convolutional Neural Network (CNN), one of the deep learning methods, was used which suggested automatic classification of CT images of lungs for early diagnosis of Covid-19 disease. In addition, k-Nearest Neighbors (k-NN) and Support Vector Machine (SVM) was used to compare the classification successes of deep learning with machine learning. Within the scope of the study, a 23-layer CNN architecture was designed and used as a classifier. Also, training and testing processes were performed for Alexnet and Mobilenetv2 CNN architectures as well. The classification results were also calculated for the case of increasing the number of images used in training for the first 23-layer CNN architecture by 5, 10, and 20 times using data augmentation methods. To reveal the effect of the change in the number of images in the training and test clusters on the results, two different training and testing processes, 2-fold and 10-fold cross-validation, were performed and the results of the study were calculated. As a result, thanks to these detailed calculations performed within the scope of the study, a comprehensive comparison of the success of the texture analysis method, machine learning, and deep learning methods in Covid-19 classification from CT images was made. The highest mean sensitivity, specificity, accuracy, F-1 score, and AUC values obtained as a result of the study were 0,9197, 0,9891, 0,9473, 0,9058, 0,9888; respectively for 2-fold cross-validation, and they were 0,9404, 0,9901, 0,9599, 0,9284, 0,9903; respectively for 10-fold cross-validation. © Springer Science+Business Media, LLC, part of Springer Nature 2020.

Entities:  

Keywords:  Convolutional neural networks (CNN); Covid-19; Deep learning; Lung CT classification; Machine learning; Texture analysis methods

Year:  2020        PMID: 33041635      PMCID: PMC7537375          DOI: 10.1007/s11042-020-09894-3

Source DB:  PubMed          Journal:  Multimed Tools Appl        ISSN: 1380-7501            Impact factor:   2.757


Introduction

In December 2019, an outbreak occurred in Wuhan of China’s Hubei province, caused by a new corona virus which is of zootonic origin and affecting acute respiratory tract in general [46]. With the spread of the virus in question in the following days, the disease began to seriously affect all countries in the world. The World Health Organization announced in March 2020 that the disease became a global pandemic. In addition, the World Health Organization named this new epidemic in question “Covid-19”. Common symptoms of the disease are fever, cough, shortness of breath, muscle pain, and weakness [36]. The disease has quite severe and negative effects on the lungs in general. In this context, many literature studies have been realised in a short time in which these effects of the disease on lungs were shown by the help of lung CT imaging. These studies reveal that besides clinical symptoms, blood and biochemical tests, lung CT imaging is an important diagnostic tool for the diagnosis of the disease as well. In a study conducted by Qin et al. [33], a clinical examination images of lung CT belonging to four cases, two men and two women with Covid-19 disease, was performed. The results of the study reveal that lung lesions due to Covid-19 show a high degree of involvement in patients with pneumonia disease. In the study carried out by Albarello et al. [2], the change in the images related to chest X-Rays and CT of lungs belonging to 2 Covid-19 cases in Italy in the disease process was investigated. In the context of the study, it was evaluated by the authors that monitoring the deterioration of lungs by radiological images is an important alternative for the early diagnosis of the disease, considering the clinical findings. In the study conducted by Lin et al. [24], the change in CT images of lungs belonging to a 61-year-old male Covid-19 patient in accordance with the progress in the disease was examined. In the scope of the study, it is reported that there is an increase in lung involvement depending on the progress of the disease. In a similar study by Li et al. [21], CT images of lung belonging to 5 Covid-19 patients, whose ages range from 10 months to 6 years were evaluated clinically. As a result of the study, it was reported that 2 of these patients had no signs of disease on CT images of lungs but 3 of them showed significant abnormalities. In the study conducted by Xu et al. [41], CT images of lungs belonging to 50 Covid-19 patients in total, of whose 9 are in mild, 28 moderate, 10 severe, and 3 critical level, were evaluated clinically. In the examinations carried out within the scope of the study, it was reported that no changes in radiological images occurred in 9 patients, while symmetrical lesions developed in 26 patients and asymmetrical lesions in 15 patients. In addition, in the discussion part of the study, it was evaluated that repeated CT scanning is a useful method for monitoring disease progress and timely treatment of Covid-19. In the study performed by Xia et al. [39], CT images of 20 Covid-19 diagnosed pediatric patients were analyzed. The results of the study reveal that all patients have sub-pleural lesions. A similar study by Chen et al. [10] was carried out by analyzing the clinical data history of 9 women who were diagnosed with Covid-19 and were also pregnant. As a result of the study, it is stated that CT images of lungs have high diagnostic value in Covid-19 diagnosis. A study of clinical data of patients diagnosed with Covid-19 was performed by Huang et al. [18]. It was reported that lung abnormalities were detected in 40 of 41 Covid-19 patients examined within the scope of the study and that there was bilateral involvement. In the study performed by Hu et al. [16], the CT images of lung belonging to 2 Covid-19 patients were evaluated. In the scope of the study, although the symptoms of the disease decreased after two days of treatment, this reduction and recovery showed inconsistency with the CT images of lungs. In the study conducted by Liu et al. [25], clinical evaluation of CT images of lungs belonging to 73 Covid-19 cases of various severities was performed. The results of the study show that all patients except the patient group of 8 percent who undergo the disease at the level of mild pneumonia had abnormal CT images of lungs. In the study conducted by Xu et al. [42], the clinical data histories of 90 Covid-19 patients were examined. From the results of the study, it was reported that Covid-19 patients had multiple patchy ground glass opacities on CT images. A similar study was carried out by Pan et al. [31], using the clinical data of 63 Covid-19 patients and abnormalities were detected in CT images of lungs belonging to the patients. As for the study conducted by Shen et al. [37], the lesion levels in CT images of lungs belonging to 44 Covid-19 patients were labeled by radiologists and computers respectively. The study results show that computerized labeling is a reliable alternative method in detecting the severity and distribution of pneumonia due to Covid-19 disease. In a study conducted by Li et al. [22], a clinical evaluation of CT images of lungs taken in the course of pneumonia due to Covid-19 disease was performed. In the study, attention was drawn to the importance of CT imaging of lungs in understanding the effect and progress of the Covid-19 disease. Clinical symptoms and CT images of lungs belonging to a 54-year-old male patient with Covid-19 disease in South Korea were investigated by Lim et al. [23]. A similar study was conducted by Cheng et al. [11] for the first Covid-19 case in Taiwan. In a review study conducted by Long and Ehrenfeld [26], it is emphasized that using artificial intelligence methods to reduce the effects of Covid-19 outbreak crisis is an essential requirement. In this context, some applications regarding the automatic classification of Covid-19 and Non-Covid-19 CT images with artificial intelligence methods were implemented. Detailed information about the methods used in these studies, number of images, training-test methods and study results are given in Table 1.
Table 1

Results of previous studies for Covid-19 and Non-Covid-19 CT classification using CT images

StudyYearNo. of imagesMethodsTest methodsResults
Han et al. [14]2020460 images (230 Covid-19 and 230 Non-Covid-19)C3D, DeCoVNet and AD3D-MIL Algorithm60% Train-20% Validation-20% TestSen: 0,968-0,979; Spe: X; Acc: 0,968-0,979; F-1 Score: 0,968-0,979; AUC: 0,982-0,990; Time: X
Ardakani et al. [4]20201020 images (510 Covid-19 and 510 Non-Covid-19)Pre-processing (Cropped / Input Image Size: 60x60) and Transfer Learning with Convolutional Neural Networks (AlexNet, VGG-16, VGG-19, SqueezeNet, GoogleNet, MobileNet-V2, ResNet-18, ResNet-50, ResNet-101, and Xception)80% Train-20% TestSen: 0,7843-1,000; Spe: 0,6863-1,000; Acc: 0,7892-0,9951; F-1 Score: X; AUC: 0,894-0,994; Time: X
Jaiswal et al. [19]20202492 images (1262 Covid-19 and 1230 Non-Covid-19)Transfer Learning with Convolutional Neural Networks (VGG16, Inception ResNet, Resnet 152V2, DenseNet201)68% Train-17% Validation-15% TestSen: 0,9206-0,9735; Spe: 0,8972-0,9621; Acc: 0,909-0,9625; F-1 Score: 0,9109-0,9629; AUC: 0,97; Time: X
Pathak et al. [32]2020852 images (413 Covid-19 and 439 Non-Covid-19)Transfer Learning with Convolutional Neural Networks (Transfer from ResNet-50 Network to a New CNN Architecture)10-fold and 50% Train- 10% Validation- 40% TestSen: 0,9146; Spe: 0,9478; Acc: 0,9302; F-1 Score: X; AUC: X; Time: X
Ouyang et al. [30]20204982 images (3389 Covid-19 and 1593 Non-Covid-19)Convolutional Neural Network (3D Resnet34) and Uniform Sampling, Size-balanced Sampling, Dual-Sampling5-fold and Train: 2186 (1094 Covid-19 and 1092 Non-Covid-19) Test: 2796 (2295 Covid-19 and 501 Non-Covid-19)Sen: 0,869; Spe: 0,901; Acc: 0,875; F-1 Score: 0,82; AUC: 0,944; Time: X
Sakagianni et al. [34]2020746 images (349 Covid-19 and 397 Non-Covid-19)Auto-ML platform (Google AutoML Cloud Vision)Train: 596 (279 Covid-19 and 317 Non-Covid-19) Validation: 73 (34 Covid-19 and 39 Non-Covid-19) Test: 77 (36 Covid-19 and 41 Non-Covid-19)Sen: 0,8831; Spe: X; Acc: X; F-1 Score: 0,8831; AUC: X Time: X
Hu et al. [17]2020300 images (150 Covid-19 and 150 Non-Covid-19l)Weakly Supervised Deep Learning5-foldSen: 0,833; Spe: 0,956; Acc: 0,906; F-1 Score: X; AUC: 0,943; Time: X
Results of previous studies for Covid-19 and Non-Covid-19 CT classification using CT images In this study, Convolutional Neural Network (CNN), one of the deep learning methods, was used which suggested automatic classification of CT images of lungs for early diagnosis of Covid-19 disease. In addition, k-Nearest Neighbors (k-NN) and Support Vector Machine (SVM) were used to compare the classification successes of deep learning with machine learning. Within the scope of the study, besides the results obtained when the images are used directly with the SVM, k-NN, and CNN classifiers, the results obtained by using Gray-Level Co-Occurrence Matrix (GLCM), Local Binary Pattern (LBP), and Local Entropy (LE) as a pre-process were found out as well. In addition, the success of the proposed method was tested separately in case that all the image data were combined as double, triple, and quadruple. Within the scope of the study, a 23-layer CNN architecture was designed and used as a classifier. In addition, training and testing processes were performed for Alexnet and Mobilenetv2 CNN architectures as well. The classification results were also calculated for the case of increasing the number of images used in training for the first 23-layer CNN architecture by 5, 10, and 20 times using data augmentation methods. In order to reveal the effect of the change in the number of images in the training and test clusters on the results, two different training and testing processes, 2-fold and 10-fold cross-validation, were performed and the results of the study were calculated. As a result, thanks to these detailed calculations performed within the scope of the study, a comprehensive comparison of the success of the texture analysis method, machine learning, and deep learning methods in Covid-19 classification from CT images was made. The results of the study indicate that analyzing CT images of lungs with the help of deep learning methods in diagnosing Covid-19 disease will speed up the diagnosis and significantly reduce the burden on healthcare workers.

Material and method

Used data

The CT images of lungs belonging to Covid-19 patients used in the study were taken from two data set created and accumulated in the form of metadata by Cohen et al. [13] and Zhao et al. [45] who then made it available to public through Github for open access. In this context, 386 CT images of lungs in total were used in the study. The sizes of the images in question are between 115×98 and 2024×1523 and vary widely. First of all, in order to clarify the area of interest on the images, framing was carried out so as to include the lung region. After this process, the dimensions of all the images were rearranged as 448×448. In the scope of the study, 1.010 Non-Covid-19 CT images of lungs were also used. The images in question were similarly taken from, LIDC-IDRI public access research database [1, 5, 6, 12], a previously published data set. The processes applied to the Covid-19 images were also applied to the Non-Covid-19 ones.

Local Binary Pattern (LBP)

Local Binary Pattern (LBP) was first introduced by Ojala et al. [27]. This method is often used to reveal local spatial structures. LBP is a sequential comparison process of a center pixel withh its neighboring pixel values. So it is a non-parametric method. The first example of this comparison was made for a 3×3 square operator. Then, operations were realized for operators of different sizes too. In (1), there is a mathematical representation of this operation. As can be seen from the (1), threshold measurement is made by realizing a comparison with neighbors of the central pixel. As a result of the threshold measurement, a LBP code with a total of 256 combinations comes out. This code is then mathematically converted to a local number, the new pixel value. A sample LBP application is shown in Fig. 1.
Fig. 1

A sample LBP application

A sample LBP application In case LBP is applied to the images, it is seen that the image sizes decrease. For example, when an LBP operation with a radius 2 is applied to an image of 224×224, an image of 220×220 is obtained. The reason for this is that this operation cannot be applied to the pixels in the starting and ending rows and columns. The diameter of the LBP used in this study is 2. To prevent this size reduction from causing any problem in the operations mentioned above, the dimensions of the image obtained after LBP operations were restored to 448×448. The use of the LBP process in the study enable us to have new images from the original ones that reflect local features. Thus, the total image feature depth was increased.

Local Entropy (LE)

Entropy can be explained as the measure of uncertainty or randomness in an image [43]. The global entropy of an image can be mathematically calculated with (2). In this formula, p is obtained from the normalized histogram of the image and is called the probability associated with the pixel value i. If local histograms are used instead of global histograms, local entropy is calculated. Local entropy is an effective feature extraction method used in texture analysis. CT lung images were included in the study to reveal the texture properties.

Gray-Level Co-Occurrence Matrix (GLCM)

Gray-Level Co-Occurrence Matrix (GLCM) is used to reveal how often pairs of pixels with a specified value and shift occur in an image [40]. The GLCM conjugate of an image can be mathematically calculated with (3). In this equation, I denote the image in size n×m, of which the GLCM is calculated, i and j are the pixel values; x and y denote the spatial position of the image and C is the co-occurrence matrix. The size of the matrix C is p×p, with p the number of different pixels in I image. For this reason, as in LBP, the re-arrangement of the dimensions is realized after the GLCM process. The GLCM, being a method frequently used for texture analysis in image processing applications, was included in the study to reveal different image features of CT lung images and to increase the variety of features. In Fig. 2, the original versions of a Covid-19 and Non-Covid-19 image and the versions obtained by applying LBP, LE, and GLCM are given.
Fig. 2

a Covid-19 b Non-Covid-19 CT image samples (Original, LBP, LE, and GLCM image; respectively)

a Covid-19 b Non-Covid-19 CT image samples (Original, LBP, LE, and GLCM image; respectively)

Data augmentation

Within the scope of the study, the effects of the increase in the number of images used in the training set of the CNN classifier on the study results were also examined. For this purpose, data were augmented using classical data augmentation methods. The number of 1.396 CT lung images (386 Covid-19 and 1.010 Non-Covid-19) was augmented by 5 times, 10 times, and 20 times using conventional data augmentation methods. In this context, data augmentation was realized by using contrast changing, brightness changing, distortion, and noise adding methods on the original images. In some steps of performing these processes, the image augmentation tools created by the works of Bloice et al. [7-9] were used. First, the pixel values related to the original image size being multiplied by 0,8 and 0,6 respectively and the contrast being changed, the second and third images were obtained. The brightness change was made by increasing the value of each pixel of the second and third images by 7 and the fourth and fifth images were created. After these processes, the number of images amounted to 5 times the previous number. Secondly, the original images were eroded by the distortion process. Later, the processes performed in the first stage were also applied to the eroded images. At the end of the second stage, the number of images amounted to 10 times the number of images obtained in the first stage. A salt and pepper noise of a density of 0,03 was added to the original image and the first nine images obtained from this image, and the number of images was increased to 20. The selection of the type of noise to be added and the parameter sizes used in the data augmentation processes was made by taking into consideration the literature studies [29] and [28] having successful results. In this context, it is possible to choose different parameter sizes and noise types and to use other data augmentation methods. By using the data augmentation methods, the amount of LBP, LE, and GLCM images as well as the original images were augmented by 5 times, 10 times, and 20 times. At this stage, data augmentation processes applied to the original images were performed using these images, instead of the original image as the source image.

Support Vector Machine (SVM)

Support Vector Machine (SVM) can be broadly described as a supervised learning approach. It is based on statistical learning theory. The foundations of SVM were demonstrated by the studies realized by Vapnik and Chervonenkis [38]. SVM can effectively perform linear and nonlinear classifications. Two-class SVM was used within the scope of the study. Radial Basis Function (RBF) was preferred as a kernel function in SVM. While the results were standardized, other parameters were used as the default (Matlab 2019a).

k-Nearest Neighbors (k-NN)

k-Nearest Neighbors (k-NN) is a non-parametric classification and regression method [3]. When classifying is made using k-NN, the object to be classified is voted on by its neighbors and assigned to the most common class among its closest neighbors. Two-class k-NN was used within the scope of the study. While the neighbor number of k-NN was set 21, other parameters were used as the default (Matlab 2019a). One of the main objectives of the study is to compare the success of machine learning with deep learning methods in Covid-19 classification. For this purpose, SVM and k-NN, which are basic machine learning methods, were included in the study as classifiers.

Convolutional Neural Network (CNN)

Deep learning is realized using too many layers in the realization of learning process. Convolutional Neural Network (CNN) is the most frequently used model of deep learning. This model has come into use widely, especially in image processing applications in recent years. CNN consist of some layers such as convolutional layer, activation function, pooling and fully connected layer. Convolutional layers are often designed to follow each other and enable us to obtain feature patterns from low-level features of images to high-level features [15]. Activation functions in CNN architecture can be defined as the functions that bring incoming inputs to a certain range or accept some of the input values while eliminating some of them again. Pooling layers, on the other hand, let the size of feature matrices to be reduced, through sampling. As for fully connected layer, it is the layer where the classification process is performed according to the features obtained through convolution, activation function and pooling. This layer works like a classic artificial neural network. Before the classification process, the conversion of feature matrices into feature vectors, that is, flattening process is performed. In this context, Fig. 3 shows the general architecture of the CNN classifier.
Fig. 3

General architecture of the CNN classifier

General architecture of the CNN classifier Within the scope of the study, a CNN architecture consisting of 23 layers in total was designed. Table 2 shows the information and parametric features of the layers of this CNN architecture. In programming processes, Matlab 2019a program was used and the function names and parameters of the layers used in the program were written directly in the layer name and parameters sections. The reason why there are four different dimensions in the first layer, the image input layer, is that experiments were made for multiple and different sized input images within the scope of the study.
Table 2

Features of CNN architecture constructed within the scope of the study

Layer numberLayer nameLayer parameters (Matlab)
1imageInputLayer[448 448 1], [448 448 2], [448 448 3], and [448 448 4]
2convolution2dLayer(3,4,‘Padding’,‘same’)
3batchNormalizationLayerdefault
4reluLayerdefault
5maxPooling2dLayer(2,‘Stride’,2)
6convolution2dLayer(3,8,‘Padding’,‘same’)
7batchNormalizationLayerdefault
8reluLayerdefault
9maxPooling2dLayer(2,‘Stride’,2)
10convolution2dLayer(3,16,‘Padding’,‘same’)
11batchNormalizationLayerdefault
12reluLayerdefault
13maxPooling2dLayer(2,‘Stride’,2)
14convolution2dLayer(3,32,‘Padding’,‘same’)
15batchNormalizationLayerdefault
16reluLayerdefault
17maxPooling2dLayer(2,‘Stride’,2)
18convolution2dLayer(3,64,‘Padding’,‘same’)
19batchNormalizationLayerdefault
20reluLayerdefault
21fullyConnectedLayer2
22softmaxLayerdefault
23classificationLayerdefault
Features of CNN architecture constructed within the scope of the study In addition to the 23-layer CNN architecture presented within the scope of the study, two more CNN architectures that were previously presented in the literature and proved to be successful in many studies, were used. In this context, Alexnet [20] and Mobilenetv2 [35] CNN architectures were also included in the study. While only the architectures of these CNN models were transferred, the previously trained coefficients were not transferred. Alexnet and Mobilenetv2 were trained from the beginning with the images used in the study. Alexnet is a CNN architecture that consists of 25 layers and has a serially arranged structure. While the input image dimensions of the original architecture of the Alexnet model are 227×227×3 the fully connected layer output size is 1000. In other words, a classification with 1000 classes is performed at the output of this architecture. Mobilenetv2, on the other hand, has an architecture consisting of 155 layers and 164 connection points that is, arranged both serially and parallelly. While the initial image dimensions of the original architecture of Mobilenetv2 model are 224×224×3, the size of the fully connected layer output size is 1000. Some modifications need to be made on Alexnet and Mobilenetv2 architectures for them to be used in the study. In this context, the number of lines and columns in the input image was preserved as the values in the original architecture. However, the size of the third dimension of the input data was rearranged as 1, 2, 3, and 4 according to the requirements of the study. In this context, the new input image sizes of the modified Alexnet architecture are 227×227×1, 227×227×2, 227×227×3, 227×227×4 and the new input image sizes of the modified Mobilenetv2 architecture are 224×224×1, 224×224×2 is 224×224×3, 224×224×4. Within the scope of the study, since a two-class classification, Covid-19 and Non-Covid-19, was performed, the fully connected layer output size was re-adjusted as 2 in both CNN architectures. For this reason, the dimensions of the images used in the operations realized using Alexnet and Mobilenetv2 are different from the first CNN architecture and machine learning classifiers. With these architectures, the images were re-sized using the imresize (Matlab 2019a) function. This issue was also taken into consideration additionally, in the interpretation of CPU run-times within the scope of the study.

The evaluation criterias for the classification results

In this study, besides the parameters: TP, TN, FN, FP and the dependent variables such as sensitivity (SEN), specificity (SPE), F-1 Score (F-1), and accuracy (ACC) which were obtained by mathematical processing of the values of the above-mentioned parameters were used in the evaluation of the results. Within the scope of the study, Receiver Operating Characteristic (ROC) analysis was made and the area (AUC) sizes under the ROC curve were compared. In this context, TP is the number of times that the actually patient data is labeled as patients as a result of classification as well. FP, on the other hand, is the number of times that non-patient data is labeled as patient in the same way. TN is the number of times that non-patient data is labeled as not patient as a result of the classification. FN, on the other hand, is the labeling of the patient data as non-patient in the same way. Sensitivity (SEN), specificity (SPE), F-1 Score (F-1), and accuracy (ACC) values calculated using these parameters are defined mathematically between (4)–(7). Within the scope of the study, SEN, SPE, F-1, and ACC values were calculated for the threshold (cut-off) value to be 0,5. ROC analysis examines the change of sensitivity (SEN) (y-axis) relative to the discrimination threshold value, that is, the cut-off value [44], in relation to the precision that is the complement of the specificity to one (1-SPE) (x-axis) in the classification processes performed in two groups. In this context, the area under the curve that reflects the change in question is called AUC. This field approaching to 1 indicates that the classification goes to perfection, while approaching 0 indicates that the classification deteriorates.

Experiments and results

Experiments

Within the scope of the study, machine learning-based and deep learning-based artificial intelligence applications were implemented for the automatic diagnosis of CT images of lung for early diagnosis of Covid-19 disease. In this study, a total of 1.396 lung CT images were used, 386 of which were belonging to patients diagnosed with Covid-19 and 1.010 to Non-Covid-19 subjects. In the study, first of all, framing process was carried out to include the lung region in order to clarify the interests of the related images. Since the images used are of very different sizes, the images were rearranged in size and the image dimensions were set to 448×448. Later, the images in question were recorded in gray-scale format and in 8 bit depth in png format. In this context, it is possible to name these processes as pre-process applications. In the second part of the study, new images were created by applying LBP, LE, and GLCM to the original images of Covid-19 patients and Non-Covid-19 patients and recorded in the same size, format and depth as the original images. In this context, since there was a decrease in image sizes after LBP and GLCM application, those images were rearranged in 448×448 dimensions. In this way, new images containing feature patterns of the relevant images were created and the features of texture were enriched. Within the scope of the study, the effects of the increase in the number of images used in the training set of the CNN classifier on the study results were also examined. In this stage of the study, data augmentation was performed by the help of changing contrast, distortion, and brightness as well as noise-adding methods described earlier. This data augmentation process was performed for the images applied LBP, LE, and GLCM as well as the original images. As in the first stage, these images were recorded in gray-scale format and 8 bit depth in png format. At the end of the second phase, the amount of data was augmented 5 times, 10 times, and 20 times, including first images on which were based derivation. In the third stage of the study, first, the original images with LBP, LE, and GLCM images separately and secondly in the form of double, triple and quadruple combinations including the original image in each group were subjected to Covid-19 and Non-covid-19 classification using SVM and k-NN classifier. In order to observe the effects of the change in the training-test image number balance on the results, two different training-test processes, that is, 2-fold cross-validation and 10-fold cross-validation, were performed and the results were calculated. In the fourth part of the study, a 23-layer CNN architecture, previously described, was developed and this architecture was used in all the experiments. In this section, the image groups created in the previous step were classified using this 23-layer CNN architecture. As in the previous stage, 2-fold cross-validation and 10-fold cross-validation were used for training-test processes in classification. Each experiment was repeated 5 times in itself in order to let the results reach stability due to the random assignment of some initial weights and parameters used in the structure of CNN.The average of the test results obtained as a result of five repetitions was given in the scope of the study. In the fifth stage of the study, the study results were calculated for the case of increasing the number of images and using them in 23-layer CNN training with the help of the data augmentation techniques detailed previously. In this context, training-testing processes were realized to increase by three different amounts: 5, 10, and 20 times. In these training-test processes, only 2-fold cross-validation was performed due to the length of processing times. The original images were divided into two, so as to be they are in the same group as the images derived from them. That is, when the original image is in the training group, it is prevented that a derived image obtained from this image is in the test group or vice versa. While the training process was performed by augmenting the data, the testing process was performed only for the original images. In the sixth part of the study, the training and testing processes were realized using the Alexnet and Mobilenetv2 CNN architecture which were detailed previously and of which the modification processes were performed. At this stage, training and testing processes were realized using 2-fold and 10-fold cross-validation as in the third and fourth stages. However, unlike these stages, instead of 448×448, the dimensions of the input images were used as 227×227 for Alexnet and 224×224 for Mobilenetv2. Within the scope of the study, the times (CPU time) needed to be able to classify an image were also measured. The experiments of this study were carried out using MATLAB 2019 (a) software running on 64 GB RAM and Intel(R) Xeon (R) CPU E5-2680 2.7 GHz (32 CPUs).

Results

Within the scope of the study, first the Covid-19/Non-Covid-19 classification was performed using the SVM classifier. The dimensions of the input images for this experiment were 448×448×1 for Original, GLCM, LBP, and LE; 448×448×2 for Original+GLCM, Original+LBP, and Original+LE; 448×448×3 for Original+GLCM+LBP, Original+GLCM+LE, and Original+LBP+LE; 448×448×4 for Original+GLCM+LBP+LE. These image matrixes were given as input to SVM matrixes after being converted to image vectors. The study results obtained in this context were given in Table 3 for 2-fold cross-validation and Table 4 for 10-fold cross-validation. The highest parameter values in Table 3 and Table 4 are stated in bold. This marking is also made in the following tables.
Table 3

Results obtained by using SVM classifier for 2-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original34244968420,88600,95840,93840,88830,97771,5857
GLCM34145981290,88340,97130,94700,90210,98151,6846
LBP85301101000,22021,00000,78440,36090,79954,1768
LE31967972380,82640,96240,92480,85870,96422,2867
Original+GLCM34244969410,88600,95940,93910,88950,98003,1503
Original+LBP33551971390,86790,96140,93550,88160,97443,6905
Original+LE34046965450,88080,95540,93480,88200,97773,4185
Original+GLCM+LBP33749971390,87310,96140,93700,88450,97845,2163
Original+GLCM+LE34442970400,89120,96040,94130,89350,98014,9756
Original+LBP+LE33551969410,86790,95940,93410,87930,97685,6864
Original+GLCM+LBP+LE34145971390,88340,96140,93980,89030,97957,0295
Table 4

Results obtained by using SVM classifier for 10-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original35432976340,91710,96630,95270,91470,98539,2526
GLCM34937989210,90410,97920,95850,92330,98749,6001
LBP94292101000,24351,00000,79080,39170,855425,4466
LE32165991190,83160,98120,93980,88430,975113,0530
Original+GLCM35135978320,90930,96830,95200,91290,986618,3973
Original+LBP34442977330,89120,96730,94630,90170,984021,4024
Original+LE34640977330,89640,96730,94770,90460,984919,8091
Original+GLCM+LBP34640977330,89640,96730,94770,90460,986030,1233
Original+GLCM+LE34838978320,90160,96830,94990,90860,986628,7369
Original+LBP+LE34244980300,88600,97030,94700,90240,985332,7798
Original+GLCM+LBP+LE34640981290,89640,97130,95060,90930,986841,9181
Results obtained by using SVM classifier for 2-fold cross-validation Results obtained by using SVM classifier for 10-fold cross-validation Within the scope of the study, secondly, the Covid-19/Non-Covid-19 classification was performed using the k-NN classifier. The dimensions of the input images for this experiment are the same as those used for the SVM classifier. These image matrixes were given as input to k-NN matrixes after being converted to image vectors. The study results obtained in this context were given in Table 5 for 2-fold cross-validation and Table 6 for 10-fold cross-validation.
Table 5

Results obtained by using k-NN classifier for 2-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original30086975350,77720,96530,91330,83220,98002,4588
GLCM267119997130,69170,98710,90540,80180,98722,5106
LBP116270929810,30050,91980,74860,39790,69872,5268
LE160226101000,41451,00000,83810,58610,98062,5013
Original+GLCM29294990200,75650,98020,91830,83670,98625,4448
Original+LBP29888979310,77200,96930,91480,83360,98305,4867
Original+LE29195992180,75390,98220,91910,83740,98675,4456
Original+GLCM+LBP286100993170,74090,98320,91620,83020,98687,9491
Original+GLCM+LE282104999110,73060,98910,91760,83060,98858,1147
Original+LBP+LE29294993170,75650,98320,92050,84030,98758,3712
Original+GLCM+LBP+LE282104998120,73060,98810,91690,82940,988811,0160
Table 6

Results obtained by using k-NN classifier for 10-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original31175980300,80570,97030,92480,85560,98093,9970
GLCM286100998120,74090,98810,91980,83630,98934,0297
LBP119267994160,30830,98420,79730,45680,69794,1210
LE179207100820,46370,99800,85030,63140,98284,0515
Original+GLCM29987993170,77460,98320,92550,85190,988510,0065
Original+LBP30878981290,79790,97130,92340,85200,98419,6495
Original+LE30185992180,77980,98220,92620,85390,98819,8070
Original+GLCM+LBP29492994160,76170,98420,92260,84480,988815,6017
Original+GLCM+LE29294999110,75650,98910,92480,84760,989515,6487
Original+LBP+LE29987997130,77460,98710,92840,85670,988815,9042
Original+GLCM+LBP+LE289971000100,74870,99010,92340,84380,990322,7328
Results obtained by using k-NN classifier for 2-fold cross-validation Results obtained by using k-NN classifier for 10-fold cross-validation Within the scope of the study, thirdly, the Covid-19/Non-Covid-19 classification was performed using the 23-layer CNN classifier. The dimensions of the input images for this experiment were 448×448×1 for Original, GLCM, LBP, and LE; 448×448×2 for Original+GLCM, Original+LBP, and Original+LE; 448×448×3 for Original+GLCM+LBP, Original+GLCM+LE, and Original+LBP+LE; 448×448×4 for Original+GLCM+LBP+LE. The study results obtained in this context were given in Table 7 for 2-fold cross-validation and Table 8 for 10-fold cross-validation. Each experiment was repeated 5 times in itself in order to let the results reach stability due to the random assignment of some initial weights and parameters used in the structure of CNN. The average of the test results obtained in five repetitions were given within the scope of the study. This is true for all experiments performed after this test.
Table 7

Results obtained by using 23-layer CNN classifier for 2-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original340,645,4952,058,00,88240,94260,92590,86810,96640,1322
GLCM339,846,2954,255,80,88030,94480,92690,86960,97270,1307
LBP301,884,2985,424,60,78190,97560,92210,84700,96760,1312
LE332,453,6979,031,00,86110,96930,93940,88710,97260,1307
Original+GLCM339,246,8944,665,40,87880,93520,91960,85810,96310,2001
Original+LBP330,455,6950,459,60,85600,94100,91750,85150,96440,1998
Original+LE340,046,0950,659,40,88080,94120,92450,86570,96480,1999
Original+GLCM+LBP341,644,4949,860,20,88500,94040,92510,86720,97140,2690
Original+GLCM+LE339,047,0948,861,20,87820,93940,92250,86240,96640,2686
Original+LBP+LE338,647,4949,460,60,87720,94000,92260,86250,96960,2693
Original+GLCM+LBP+LE342,243,8950,859,20,88650,94140,92620,86930,96860,3391
Table 8

Results obtained by using 23-layer CNN classifier for 10-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original357,029,0978,431,60,92490,96870,95660,92170,98451,1252
GLCM349,436,6986,623,40,90520,97680,95700,92090,99021,1235
LBP322,463,6996,413,60,83520,98650,94470,89310,98141,1242
LE340,445,6991,818,20,88190,98200,95430,91430,98491,1236
Original+GLCM354,032,0974,235,80,91710,96460,95140,91260,97891,7136
Original+LBP349,836,2979,230,80,90620,96950,95200,91260,98541,7178
Original+LE349,436,6977,033,00,90520,96730,95010,90940,98181,7174
Original+GLCM+LBP350,435,6980,829,20,90780,97110,95360,91530,98732,3014
Original+GLCM+LE350,635,4980,229,80,90830,97050,95330,91490,98432,3003
Original+LBP+LE349,436,6983,426,60,90520,97370,95470,91710,98732,3034
Original+GLCM+LBP+LE348,637,4981,428,60,90310,97170,95270,91360,98492,8945
Results obtained by using 23-layer CNN classifier for 2-fold cross-validation Results obtained by using 23-layer CNN classifier for 10-fold cross-validation Within the scope of the study, fourthly, the Covid-19/Non-Covid-19 classification was performed using the 23-layer CNN classifier and data augmentation (×5, ×10, and ×20). The dimensions of the input images for this experiment were 448×448×1 for Original, GLCM, LBP, and LE; 448×448×2 for Original+GLCM, Original+LBP, and Original+LE; 448×448×3 for Original+GLCM+LBP, Original+GLCM+LE, and Original+LBP+LE; 448×448×4 for Original+GLCM+LBP+LE. In this experiment, training and testing processes were performed using only 2-fold cross-validation due to the increase in the time to obtain results in parallel with the increase in the training set. The results obtained in the experiments were given in Table 9 for increasing the number of training images by 5 times, in Table 10 for increasing by 10 times, and in Table 11 for 20 times.
Table 9

Results obtained by using 23-layer CNN classifier and data augmentation (5 times) for 2-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original333,652,4934,875,20,86420,92550,90860,83970,95450,6499
GLCM330,056,0951,858,20,85490,94240,91820,85260,95980,6438
LBP275,8110,2977,232,80,71450,96750,89760,79400,95800,6448
LE326,659,4974,635,40,84610,96500,93210,87340,97030,6479
Original+GLCM339,047,0940,070,00,87820,93070,91620,85280,96040,9850
Original+LBP338,048,0953,456,60,87560,94400,92510,86590,96850,9868
Original+LE338,048,0949,061,00,87560,93960,92190,86100,96960,9858
Original+GLCM+LBP338,847,2951,258,80,87770,94180,92410,86470,96681,3134
Original+GLCM+LE331,654,4945,864,20,85910,93640,91500,84840,96031,3129
Original+LBP+LE341,844,2959,450,60,88550,94990,93210,87820,97361,3140
Original+GLCM+LBP+LE339,047,0957,252,80,87820,94770,92850,87170,96911,6378
Table 10

Results obtained by using 23-layer CNN classifier and data augmentation (10 times) for 2-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original334,251,8938,671,40,86580,92930,91170,84440,95651,2964
GLCM327,658,4956,453,60,84870,94690,91980,85420,96171,2972
LBP291,694,4979,630,40,75540,96990,91060,82350,95971,2998
LE330,655,4978,631,40,85650,96890,93780,88400,97091,3060
Original+GLCM336,249,8948,661,40,87100,93920,92030,85800,96251,9531
Original+LBP335,051,0957,053,00,86790,94750,92550,86560,96951,9568
Original+LE334,651,4951,658,40,86680,94220,92130,85910,96361,9528
Original+GLCM+LBP342,044,0958,251,80,88600,94870,93140,87710,97232,6482
Original+GLCM+LE345,840,2948,661,40,89590,93920,92720,87190,97072,6338
Original+LBP+LE339,846,2966,243,80,88030,95660,93550,88300,97242,6160
Original+GLCM+LBP+LE339,246,8974,435,60,87880,96480,94100,89170,97803,5341
Table 11

Results obtained by using 23-layer CNN classifier and data augmentation (20 times) for 2-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original329,656,4939,270,80,85390,92990,90890,83820,95662,6773
GLCM332,653,4934,076,00,86170,92480,90730,83720,95662,6038
LBP299,886,2971,039,00,77670,96140,91030,82720,96342,6043
LE333,053,0974,635,40,86270,96500,93670,88280,97312,6017
Original+GLCM336,649,4945,264,80,87200,93580,91820,85490,96484,1130
Original+LBP335,650,4960,050,00,86940,95050,92810,86990,97003,9790
Original+LE340,046,0959,051,00,88080,94950,93050,87520,96684,0007
Original+GLCM+LBP335,850,2965,045,00,86990,95540,93180,87590,974512,6129
Original+GLCM+LE338,247,8966,443,60,87620,95680,93450,88090,974312,6575
Original+LBP+LE342,443,6961,448,60,88700,95190,93400,88130,975311,4009
Original+GLCM+LBP+LE341,844,2960,449,60,88550,95090,93280,87930,983516,3373
Results obtained by using 23-layer CNN classifier and data augmentation (5 times) for 2-fold cross-validation Results obtained by using 23-layer CNN classifier and data augmentation (10 times) for 2-fold cross-validation Results obtained by using 23-layer CNN classifier and data augmentation (20 times) for 2-fold cross-validation Within the scope of the study, fifthly, the Covid-19/Non-Covid-19 classification was performed using the Alexnet CNN classifier. The dimensions of the input images for this experiment were 448×448×1 for Original, GLCM, LBP, and LE; 448×448×2 for Original+GLCM, Original+LBP, and Original+LE; 448×448×3 for Original+GLCM+LBP, Original+GLCM+LE, and Original+LBP+LE; 448×448×4 for Original+GLCM+LBP+LE. The study results obtained in this context were given in Table 12 for 2-fold cross-validation and Table 13 for 10-fold cross-validation.
Table 12

Results obtained by using Alexnet CNN classifier for 2-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original334,451,6932,277,80,86630,92300,90730,83810,95180,0557
GLCM336,649,4929,280,80,87200,92000,90670,83810,95580,0516
LBP316,869,2944,665,40,82070,93520,90360,82370,94820,0517
LE348,637,4971,638,40,90310,96200,94570,90200,97670,0518
Original+GLCM342,243,8934,475,60,88650,92510,91450,85140,95570,0704
Original+LBP338,447,6924,285,80,87670,91500,90440,83550,94800,0700
Original+LE346,439,6928,881,20,89740,91960,91350,85180,95860,0687
Original+GLCM+LBP346,040,0918,891,20,89640,90970,90600,84060,95450,0885
Original+GLCM+LE348,038,0936,473,60,90160,92710,92010,86220,96030,0882
Original+LBP+LE349,037,0920,289,80,90410,91110,90920,84660,95380,0881
Original+GLCM+LBP+LE343,043,0914,096,00,88860,90500,90040,83180,94700,1062
Table 13

Results obtained by using Alexnet CNN classifier for 10-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original350,635,4965,644,40,90830,95600,94280,89780,97380,4070
GLCM345,240,8974,435,60,89430,96480,94530,90030,97740,4026
LBP338,247,8970,040,00,87620,96040,93710,88510,97310,4027
LE347,238,8991,218,80,89950,98140,95870,92340,98200,4003
Original+GLCM353,232,8970,040,00,91500,96040,94790,90660,97380,5537
Original+LBP356,030,0964,645,40,92230,95500,94600,90420,97400,5550
Original+LE353,232,8971,438,60,91500,96180,94890,90820,97880,5542
Original+GLCM+LBP356,429,6959,650,40,92330,95010,94270,89910,97230,7096
Original+GLCM+LE351,634,4967,043,00,91090,95740,94460,90080,97290,7082
Original+LBP+LE359,027,0962,048,00,93010,95250,94630,90540,97200,7082
Original+GLCM+LBP+LE354,831,2956,853,20,91920,94730,93950,89370,96760,8628
Results obtained by using Alexnet CNN classifier for 2-fold cross-validation Results obtained by using Alexnet CNN classifier for 10-fold cross-validation Within the scope of the study, sixthly, the Covid-19/Non-Covid-19 classification was performed using the Mobilenetv2 CNN classifier. The dimensions of the input images for this experiment were 448×448×1 for Original, GLCM, LBP, and LE; 448×448×2 for Original+GLCM, Original+LBP, and Original+LE; 448×448×3 for Original+GLCM+LBP, Original+GLCM+LE, and Original+LBP+LE; 448×448×4 for Original+GLCM+LBP+LE. The study results obtained in this context were given in Table 14 for 2-fold cross-validation and Table 15 for 10-fold cross-validation.
Table 14

Results obtained by using Mobilenetv2 CNN classifier for 2-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original355,031,0964,845,20,91970,95520,94540,90310,97590,7444
GLCM336,449,6967,842,20,87150,95820,93420,87990,97750,7382
LBP262,4123,6952,657,40,67980,94320,87030,74310,92170,7395
LE270,0116,0978,831,20,69950,96910,89460,78550,93790,7366
Original+GLCM353,632,4968,841,20,91610,95920,94730,90580,97830,7595
Original+LBP319,666,4960,249,80,82800,95070,91680,84610,96040,7516
Original+LE333,852,2962,048,00,86480,95250,92820,86950,96280,7622
Original+GLCM+LBP326,259,8959,250,80,84510,94970,92080,85510,96930,7743
Original+GLCM+LE337,448,6965,045,00,87410,95540,93300,87820,96270,7790
Original+LBP+LE323,262,8972,637,40,83730,96300,92820,86560,96660,7725
Original+GLCM+LBP+LE323,462,6967,043,00,83780,95740,92440,85960,96430,7987
Table 15

Results obtained by using Mobilenetv2 CNN classifier for 10-fold cross-validation

MethodTPFNTNFPSENSPEACCF-1AUCCPU
Time
Original363,023,0977,033,00,94040,96730,95990,92840,98336,6215
GLCM352,233,8983,426,60,91240,97370,95670,92100,98606,6210
LBP285,8100,2967,043,00,74040,95740,89740,79960,94526,5895
LE307,878,2988,022,00,79740,97820,92820,86000,96776,6118
Original+GLCM360,026,0974,036,00,93260,96440,95560,92070,98586,7781
Original+LBP349,436,6976,034,00,90520,96630,94940,90820,97946,8004
Original+LE348,437,6972,637,40,90260,96300,94630,90280,97896,7878
Original+GLCM+LBP351,634,4977,232,80,91090,96750,95190,91270,98416,9414
Original+GLCM+LE358,227,8975,234,80,92800,96550,95520,91960,98166,9408
Original+LBP+LE341,644,4972,637,40,88500,96300,94140,89310,97456,9451
Original+GLCM+LBP+LE346,439,6976,233,80,89740,96650,94740,90410,98387,1508
Results obtained by using Mobilenetv2 CNN classifier for 2-fold cross-validation Results obtained by using Mobilenetv2 CNN classifier for 10-fold cross-validation

Conclusion

Within the scope of the study, the results of the study are given with a total of 13 Tables (Tables 3–15) in which there are experimental results for 11 different input image in each table. The evaluation of the best results obtained in these experiments has carried out in this section. When Tables 3 and 4, which includes the results obtained by using the SVM classifier, are examined within the scope of the study, the highest mean sensitivity, specificity, accuracy, F-1 score, and AUC values were found to be 0,8912, 1,0000, 0,9470, 0,9021, 0,9815; respectively for 2-fold cross-validation, 0,9171, 1,0000, 0,9585, 0,9233, 0,9874; respectively for 10-fold cross-validation. It was observed that although the highest mean specificity value was obtained as 1,0000, there were serious worsenings in sensitivity, accuracy, F-1 score, and AUC values in experiments where this value was obtained. The second highest specificity obtained using the SVM classifier is 0,9713 for 2-fold cross-validation and 0,9812 for 10-fold cross-validation. In obtaining these highest average values mentioned, it is seen that the combinations of input images such as GLCM and Original+GLCM+LE for 2-fold cross-validation, the Original, GLCM, and LE were used for 10-fold cross-validation. When the processing times are examined, it is understood that results can be obtained for an image at times ranging from 1,5857 to 7,0295 seconds for 2-fold cross-validation and from 9,2526 to 41,9181 seconds for 10-fold cross-validation, in direct proportion to the image data sizes. When Tables 5 and 6, which includes the results obtained by using the k-NN classifier, are examined within the scope of the study, the highest mean sensitivity, specificity, accuracy, F-1 score, and AUC values were found to be 0,7772, 1,0000, 0,9205, 0,8403, 0,9888; respectively for 2-fold cross-validation, 0,8057, 0,9980, 0,9284, 0,8567, 0,9903; respectively for 10-fold cross-validation. As is with the SVM classifier, although the highest mean specificity values were obtained for the k-NN classifier as 1,0000 and 0,9980 as well, serious worsenings occur in the sensitivity, accuracy, F-1 score, and AUC values in the experiments where these values were obtained. The second highest specificity obtained using the k-NN classifier is 0,9891 for 2-fold cross-validation and 0,9901 for 10-fold cross-validation. In obtaining these highest average values mentioned, it is seen that the combinations of input images such as Original, Original+GLCM+LE, Original+LBP+LE, and Original+GLCM+LBP+LE for 2-fold cross-validation, the Original, Original+LBP+LE, and Original+GLCM+LBP+LE were used for 10-fold cross-validation. When the processing times are examined, it is understood that results can be obtained for an image at times ranging from 2,4588 to 11,0160 seconds for 2-fold cross-validation and from 3,9970 to 22,7328 seconds for 10-fold cross-validation, in direct proportion to the image data sizes. When Tables 7 and 8, which includes the results obtained by using the 23-layer CNN classifier, are examined within the scope of the study, the highest mean sensitivity, specificity, accuracy, F-1 score, and AUC values were found to be 0,8865, 0,9756, 0,9394, 0,8871, 0,9727; respectively for 2-fold cross-validation, 0,9249, 0,9865, 0,9570, 0,9217, 0,9902; respectively for 10-fold cross-validation. In obtaining these highest average values mentioned, it is seen that the combinations of input images such as GLCM, LBP, LE, and Original+GLCM+LBP+LE for 2-fold cross-validation, the Original, GLCM, and LBP were used for 10-fold cross-validation. When the processing times are examined, it is understood that results can be obtained for an image at times ranging from 0,1307 to 0,3391 seconds for 2-fold cross-validation and from 1,1235 to 2,8945 seconds for 10-fold cross-validation, in direct proportion to the image data sizes. The results obtained in the experiments using a 23-layer CNN classifier and data augmentation are shown between Tables 9 and 11 for 2-fold cross-validation. When these tables are examined, the highest mean sensitivity, specificity, accuracy, F-1 score, and AUC values are 0,8855, 0,9675, 0,9321, 0,8782, 0,9736; respectively for 5 times increment, 0,8959, 0,9699, 0,9410, 0,8917, 0,9780; respectively for 10 times increment and 0,8870, 0,9650, 0,9367, 0,8828, 0,9835; respectively for 20 times increment. It is seen that, in obtaining these highest mean values, the input image combinations of LBP, LE, and Original+LBP+LE for 5 times increment; LBP, Original+GLCM+LE, and Original+GLCM+LBP+LE for 10 times increment and LE, Original+LBP+LE, and Original+GLCM+LBP+LE for 20 times increment were used. When the processing times are examined, it is understood that results can be obtained for an image at times ranging from 0,6438 to 1,6378 seconds for 5 times increment, from 1,2964 to 3,5341 seconds for 10 times increment, and from 2,6017 to 16,3373 seconds for 20 times increment, in direct proportion to the image data sizes. When Tables 12 and 13, which includes the results obtained by using the Alexnet CNN classifier, are examined within the scope of the study, the highest mean sensitivity, specificity, accuracy, F-1 score, and AUC values were found to be 0,9041, 0,9620, 0,9457, 0,9020, 0,9767; respectively for 2-fold cross-validation, 0,9301, 0,9814, 0,9587, 0,9234, 0,9820; respectively for 10-fold cross-validation. In obtaining these highest average values mentioned, it is seen that the combinations of input images such as LE and Original+LBP+LE for 2-fold cross-validation and 10-fold cross-validation. When the processing times are examined, it is understood that results can be obtained for an image at times ranging from 0,0516 to 0,1062 seconds for 2-fold cross-validation and from 0,4003 to 0,8628 seconds for 10-fold cross-validation, in direct proportion to the image data sizes. When Tables 14 and 15, which includes the results obtained by using the Mobilenetv2 CNN classifier, are examined within the scope of the study, the highest mean sensitivity, specificity, accuracy, F-1 score, and AUC values were found to be 0,9197, 0,9691, 0,9473, 0,9058, 0,9783 respectively for 2-fold cross-validation, 0,9404, 0,9782, 0,9599, 0,9284, 0,9860 respectively, for 10-fold cross-validation. In obtaining these highest average values mentioned, it is seen that the combinations of input images such as Original, LE, and Original+GLCM for 2-fold cross-validation, the Original, GLCM and LE were used for 10-fold cross-validation. When the processing times are examined, it is understood that results can be obtained for an image at times ranging from 0,7366 to 0,7987 seconds for 2-fold cross-validation and from 6,5895 to 7,1508 seconds for 10-fold cross-validation, in direct proportion to the image data sizes. Table 16 contains the comparison of the results obtained in the literature studies, of which a piece of summary information was given in the introduction section, with the results obtained within the scope of the study. As can be seen from Table 16, the highest mean sensitivity, specificity, accuracy, F-1 score, and AUC values obtained within the scope of the study were 0,9197, 0,9891 (the second highest), 0,9473, 0,9058, 0,9888 respectively, for 2-fold cross-validation and 0,9404, 0,9901 (the second highest), 0,9599, 0,9284, 0,9903 respectively for 10-fold cross-validation.
Table 16

Comparison of the results obtained within the scope of the study with previous studies

StudySENSPEACCF-1AUC
Han et al.[20]0,968-0,979X0,968-0,9790,968-0,9790,982-0,990
Ardakani et al. [21]0,7843-1,0000,6863-1,0000,7892-0,9951X0,894-0,994
Jaiswal et al. [22]0,9206-0,97350,8972-0,96210,909-0,96250,9109-0,96290,97
Pathak et al. [23]0,91460,94780,9302XX
Ouyang et al. [24]0,8690,9010,8750,8200,944
Sakagianni et al. [25]0,8831XX0,8831X
Hu et al. [26]0,83300,95600,9060X0,9430
Our Study (2- fold)0,91970,98910,94730,90580,9888
Our Study (10- fold)0,94040,99010,95990,92840,9903
Comparison of the results obtained within the scope of the study with previous studies

Discussion

Important results were obtained within the scope of this study, which recommends automatic classification of CT images of lungs for early diagnosis of Covid-19 disease and benefits from texture analysis methods, machine learning and deep learning. Within the scope of the study, the classification success of texture analysis methods, machine learning and deep learning methods were compared comprehensively for many method combinations. Within the scope of the study, if the SVM and k-NN classifiers are used as machine learning methods are compared, it is seen that the SVM classifier is more successful than the k-NN classifier. As for the classification time per image, it is understood that the SVM classifier for 2-fold cross-validation and the k-NN classifier for 10-fold cross-validation produce faster classification results. When the results obtained by using only the original images for SVM and k-NN classifiers are compared with the highest classification results obtained by adding new texture feature images, it is understood that significant improvements were achieved in many comparison parameters. Using new texture feature images, improvements were observed in the sensitivity, specificity, accuracy, F-1 score, and AUC parameters for SVM classifier and 2-fold cross-validation by 0,5181%, 1,2871% (with reference to the second-highest result), 0,8596%, 1,3805%, 0,3817% respectively. These improvements were; 1,4851% (with reference to the second-highest result), 0,5731%, 0,8552%, 0,2096% respectively, in the specificity, accuracy, F-1 score, and AUC parameters for the SVM classifier and 10-fold cross-validation. The highest sensitivity values for 2-fold cross-validation and 10-fold cross-validation using the k-NN classifier were obtained using the original images. However, it is seen that the use of new texture feature images reflected positively on other comparison parameters. Using the new texture feature images, it is seen that improvements in the specificity, accuracy, F-1 score, and AUC parameters for the k-NN classifier and 2-fold cross-validation by 2,3762% (with reference to the second-highest result), 0,7163%, 0,8110%, 0,8871% respectively and 1,9802% (with reference to the second highest result), 0,3582%, 0,1163%, 0,9370% respectively for 10-fold cross-validation were achieved. In this context, it is understood that the direct use of GLCM and LE texture feature extraction methods in the improvement of the classification results using the SVM classifier, and the use of LBP and LE tissue feature extraction methods in combination with the original images in the improvement of the classification results using the k-NN classifier will provide the most effective benefit. Within the scope of the study, the effect of increasing the number of images used in education by using data augmentation on the classification results for the CNN classifier was examined. Because of the significant increases in training load and classification time due to data augmentation, the experiments were realized only for the use of 23-layer CNN and 2-fold cross-validation. The results of the study show that data augmentation has partial positive contributions on the classification. However, it is understood that it causes the same rates of increases in classification times as well. Also, when the results obtained by using only the original images for 23-layer CNN and data augmentation are compared with the highest classification results obtained by adding new texture feature images, it is understood that significant improvements were achieved in many comparison parameters. If a comparison of the CNN architectures used within the scope of the study is required, it is seen that the most successful results are obtained by using Mobilenetv2. In terms of classification times though, it is understood that the fastest results are achieved by using Alexnet and 23-layer CNN. There is a sevenfold difference between Alexnet and Mobilenetv2 in terms of the speed of achieving results. Similar to the cases of SVM and k-NN, it is understood that the inclusion of new texture feature images into the classification has positive effects on the classification results for CNN classifiers, as well. Some conclusions were obtained for CNN classifiers, by comparing the results obtained by using only the original images with the highest classification results obtained by adding new texture feature images. In this context, using new tissue feature images, it was seen that improvements were achieved in the sensitivity, specificity, accuracy, F-1 score, and AUC parameters by 0,4145%, 3,3069%, 1,3467%, 1,8929%, 0,6367% respectively, for 23-layer CNN classifier and 2-fold cross-validation. The highest sensitivity and F-1 score values for 10-fold cross-validation using a 23-layer CNN classifier were obtained using the original images. However, it is seen that the use of new texture feature images reflected positively on other comparison parameters. In this context, it is seen that by using new texture feature images, improvements in the specificity, accuracy, and AUC parameters for a 23-layer CNN classifier and 10-fold cross-validation by 1,7822%, 0,0430%, 0,5768%, respectively, were achieved. For the Alexnet classifier, those improvements were even higher. Using new tissue feature images, improvements in the sensitivity, specificity, accuracy, F-1 score, and AUC parameters for the Alexnet classifier and 2-fold cross-validation by 3,7824%, 3,9010%, 3,8395%, 6,3918%, 2,4969% respectively, and for 10-fold cross-validation, by 2,1762%, 2,5347%, 1,5903%, 2,5606%, and 0,8262% respectively, were achieved. The highest sensitivity value for 2-fold cross-validation using the Mobilenetv2 classifier was obtained using the original images. However, it is seen that the use of new texture feature images reflected positively on other comparison parameters. In this context, using new texture feature images, it is seen that improvements in the specificity, accuracy, F-1 score, and AUC parameters for Mobilenetv2 classifier and 2-fold cross-validation by 1,3861%, 0,1862%, 0,2644%, 0,2356%, respectively were achieved. The highest sensitivity, accuracy, and F-1 score value for 10-fold cross-validation using the Mobilenetv2 classifier was obtained using the original images. However, it is understood that using the new texture feature images provided improvements in the specificity and AUC parameters for the Mobilenetv2 classifier and 10-fold cross-validation by 1,0891% and 0,2682%, respectively. If deep learning classifiers are compared with machine learning classifiers, it is seen that deep learning classifiers are more successful than machine learning classifiers. In addition, deep learning classifiers produce results at least 5 times faster than machine learning classifiers. When the results of the study obtained within the scope of the study and summary information of which is given in Table 16 are compared with the results obtained in previous studies in the literature, it is seen that successful results were obtained. In this context, it is understood that higher results were obtained than the results obtained in the studies performed by Pathak et al. [23], Ouyang et al. [24], Sakagianni et al. [25], and Hu et al. [26]. Study results are generally equivalent to the results obtained in the study by Jaiswal et al. [22] and behind the studies performed by Han et al. [20] and Ardakani et al. [21]. However, for making a fairer comparison, it should be taken into account that in the study conducted by Ardakani et al. [21], the determination of 60×60 ROI areas on the image was done manually and dense areas were marked. Indeed, although similar transfer learning approaches were used by Jaiswal et al. [22], poorer results were obtained because the image was used as a whole. In our study, images were classified as a whole as well. One of the most important implications of this study is that if the images obtained by texture feature methods are used directly or together with the original ones, significant improvements can be achieved in the study results. This situation is applicable to other studies with high results. In this way, it is considered that the results obtained in other studies can be enhanced even more. The results of the study indicate that analyzing CT images of lungs by the help of deep learning methods in diagnosing Covid-19 disease will speed up the diagnosis and significantly reduce the burden on healthcare workers. In this context, it is critical to increase the number of radiological and clinical data of Covid-19 patients and to make them available to researchers through open access in order to improve these studies and obtain better results. In the studies to be carried out after this stage, it is aimed to make automatic classification of X-Ray images of chest, which is an important diagnostic tool such as CT images of lung, by the help of methods based on deep learning, in Covid-19 disease diagnosis. In addition, it will be tried to increase the success of study by making use of complex-valued CNN and transfer learning approaches. Finally, it is evaluated that another important study topic is the realization of pipeline methods that will provide better results by combining the results obtained using different extraction property methods.
  29 in total

1.  The Cancer Imaging Archive (TCIA): maintaining and operating a public information repository.

Authors:  Kenneth Clark; Bruce Vendt; Kirk Smith; John Freymann; Justin Kirby; Paul Koppel; Stephen Moore; Stanley Phillips; David Maffitt; Michael Pringle; Lawrence Tarbox; Fred Prior
Journal:  J Digit Imaging       Date:  2013-12       Impact factor: 4.056

Review 2.  A Review of Coronavirus Disease-2019 (COVID-19).

Authors:  Tanu Singhal
Journal:  Indian J Pediatr       Date:  2020-03-13       Impact factor: 1.967

3.  Imaging and clinical features of patients with 2019 novel coronavirus SARS-CoV-2.

Authors:  Xi Xu; Chengcheng Yu; Jing Qu; Lieguang Zhang; Songfeng Jiang; Deyang Huang; Bihua Chen; Zhiping Zhang; Wanhua Guan; Zhoukun Ling; Rui Jiang; Tianli Hu; Yan Ding; Lin Lin; Qingxin Gan; Liangping Luo; Xiaoping Tang; Jinxin Liu
Journal:  Eur J Nucl Med Mol Imaging       Date:  2020-02-28       Impact factor: 9.236

4.  Clinical and CT features in pediatric patients with COVID-19 infection: Different points from adults.

Authors:  Wei Xia; Jianbo Shao; Yu Guo; Xuehua Peng; Zhen Li; Daoyu Hu
Journal:  Pediatr Pulmonol       Date:  2020-03-05

5.  The Role of Augmented Intelligence (AI) in Detecting and Preventing the Spread of Novel Coronavirus.

Authors:  Justin B Long; Jesse M Ehrenfeld
Journal:  J Med Syst       Date:  2020-02-04       Impact factor: 4.460

6.  Clinical features of patients infected with 2019 novel coronavirus in Wuhan, China.

Authors:  Chaolin Huang; Yeming Wang; Xingwang Li; Lili Ren; Jianping Zhao; Yi Hu; Li Zhang; Guohui Fan; Jiuyang Xu; Xiaoying Gu; Zhenshun Cheng; Ting Yu; Jiaan Xia; Yuan Wei; Wenjuan Wu; Xuelei Xie; Wen Yin; Hui Li; Min Liu; Yan Xiao; Hong Gao; Li Guo; Jungang Xie; Guangfa Wang; Rongmeng Jiang; Zhancheng Gao; Qi Jin; Jianwei Wang; Bin Cao
Journal:  Lancet       Date:  2020-01-24       Impact factor: 79.321

7.  Application of deep learning technique to manage COVID-19 in routine clinical practice using CT images: Results of 10 convolutional neural networks.

Authors:  Ali Abbasian Ardakani; Alireza Rajabzadeh Kanafi; U Rajendra Acharya; Nazanin Khadem; Afshin Mohammadi
Journal:  Comput Biol Med       Date:  2020-04-30       Impact factor: 4.589

8.  Case of the Index Patient Who Caused Tertiary Transmission of COVID-19 Infection in Korea: the Application of Lopinavir/Ritonavir for the Treatment of COVID-19 Infected Pneumonia Monitored by Quantitative RT-PCR.

Authors:  Jaegyun Lim; Seunghyun Jeon; Hyun Young Shin; Moon Jung Kim; Yu Min Seong; Wang Jun Lee; Kang Won Choe; Yu Min Kang; Baeckseung Lee; Sang Joon Park
Journal:  J Korean Med Sci       Date:  2020-02-17       Impact factor: 2.153

9.  A Novel Coronavirus from Patients with Pneumonia in China, 2019.

Authors:  Na Zhu; Dingyu Zhang; Wenling Wang; Xingwang Li; Bo Yang; Jingdong Song; Xiang Zhao; Baoying Huang; Weifeng Shi; Roujian Lu; Peihua Niu; Faxian Zhan; Xuejun Ma; Dayan Wang; Wenbo Xu; Guizhen Wu; George F Gao; Wenjie Tan
Journal:  N Engl J Med       Date:  2020-01-24       Impact factor: 91.245

Review 10.  Coronavirus Disease (COVID-19): Spectrum of CT Findings and Temporal Progression of the Disease.

Authors:  Mingzhi Li; Pinggui Lei; Bingliang Zeng; Zongliang Li; Peng Yu; Bing Fan; Chuanhong Wang; Zicong Li; Jian Zhou; Shaobo Hu; Hao Liu
Journal:  Acad Radiol       Date:  2020-03-20       Impact factor: 3.173

View more
  13 in total

1.  Deep learning-based important weights-only transfer learning approach for COVID-19 CT-scan classification.

Authors:  Tejalal Choudhary; Shubham Gujar; Anurag Goswami; Vipul Mishra; Tapas Badal
Journal:  Appl Intell (Dordr)       Date:  2022-07-18       Impact factor: 5.019

2.  Automated diagnosis of COVID stages from lung CT images using statistical features in 2-dimensional flexible analytic wavelet transform.

Authors:  Rajneesh Kumar Patel; Manish Kashyap
Journal:  Biocybern Biomed Eng       Date:  2022-07-01       Impact factor: 5.687

3.  A new deep learning pipeline to detect Covid-19 on chest X-ray images using local binary pattern, dual tree complex wavelet transform and convolutional neural networks.

Authors:  Huseyin Yasar; Murat Ceylan
Journal:  Appl Intell (Dordr)       Date:  2020-11-04       Impact factor: 5.086

4.  COVID-19 Detection System Using Chest CT Images and Multiple Kernels-Extreme Learning Machine Based on Deep Neural Network.

Authors:  M Turkoglu
Journal:  Ing Rech Biomed       Date:  2021-01-27

5.  A novel study for automatic two-class COVID-19 diagnosis (between COVID-19 and Healthy, Pneumonia) on X-ray images using texture analysis and 2-D/3-D convolutional neural networks.

Authors:  Huseyin Yaşar; Murat Ceylan
Journal:  Multimed Syst       Date:  2022-01-29       Impact factor: 2.603

6.  Detecting COVID-19 Pneumonia over Fuzzy Image Enhancement on Computed Tomography Images.

Authors:  Ali Alzahrani; Md Al-Amin Bhuiyan; Fahima Akhter
Journal:  Comput Math Methods Med       Date:  2022-01-18       Impact factor: 2.238

7.  Systematic Review of Artificial Intelligence in Acute Respiratory Distress Syndrome for COVID-19 Lung Patients: A Biomedical Imaging Perspective.

Authors:  Jasjit S Suri; Sushant Agarwal; Suneet Gupta; Anudeep Puvvula; Klaudija Viskovic; Neha Suri; Azra Alizad; Ayman El-Baz; Luca Saba; Mostafa Fatemi; D Subbaram Naidu
Journal:  IEEE J Biomed Health Inform       Date:  2021-11-05       Impact factor: 5.772

8.  A privacy-aware method for COVID-19 detection in chest CT images using lightweight deep conventional neural network and blockchain.

Authors:  Arash Heidari; Shiva Toumaj; Nima Jafari Navimipour; Mehmet Unal
Journal:  Comput Biol Med       Date:  2022-03-28       Impact factor: 6.698

9.  COVID-index: A texture-based approach to classifying lung lesions based on CT images.

Authors:  Vitória de Carvalho Brito; Patrick Ryan Sales Dos Santos; Nonato Rodrigues de Sales Carvalho; Antonio Oseas de Carvalho Filho
Journal:  Pattern Recognit       Date:  2021-06-06       Impact factor: 7.740

10.  Deep Learning-Based Approaches to Improve Classification Parameters for Diagnosing COVID-19 from CT Images.

Authors:  Huseyin Yasar; Murat Ceylan
Journal:  Cognit Comput       Date:  2021-07-15       Impact factor: 4.890

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.