Literature DB >> 34764560

A new deep learning pipeline to detect Covid-19 on chest X-ray images using local binary pattern, dual tree complex wavelet transform and convolutional neural networks.

Huseyin Yasar1, Murat Ceylan2.   

Abstract

In this study, which aims at early diagnosis of Covid-19 disease using X-ray images, the deep-learning approach, a state-of-the-art artificial intelligence method, was used, and automatic classification of images was performed using convolutional neural networks (CNN). In the first training-test data set used in the study, there were 230 X-ray images, of which 150 were Covid-19 and 80 were non-Covid-19, while in the second training-test data set there were 476 X-ray images, of which 150 were Covid-19 and 326 were non-Covid-19. Thus, classification results have been provided for two data sets, containing predominantly Covid-19 images and predominantly non-Covid-19 images, respectively. In the study, a 23-layer CNN architecture and a 54-layer CNN architecture were developed. Within the scope of the study, the results were obtained using chest X-ray images directly in the training-test procedures and the sub-band images obtained by applying dual tree complex wavelet transform (DT-CWT) to the above-mentioned images. The same experiments were repeated using images obtained by applying local binary pattern (LBP) to the chest X-ray images. Within the scope of the study, four new result generation pipeline algorithms having been put forward additionally, it was ensured that the experimental results were combined and the success of the study was improved. In the experiments carried out in this study, the training sessions were carried out using the k-fold cross validation method. Here the k value was chosen as 23 for the first and second training-test data sets. Considering the average highest results of the experiments performed within the scope of the study, the values of sensitivity, specificity, accuracy, F-1 score, and area under the receiver operating characteristic curve (AUC) for the first training-test data set were 0,9947, 0,9800, 0,9843, 0,9881 and 0,9990 respectively; while for the second training-test data set, they were 0,9920, 0,9939, 0,9891, 0,9828 and 0,9991; respectively. Within the scope of the study, finally, all the images were combined and the training and testing processes were repeated for a total of 556 X-ray images comprising 150 Covid-19 images and 406 non-Covid-19 images, by applying 2-fold cross. In this context, the average highest values of sensitivity, specificity, accuracy, F-1 score, and AUC for this last training-test data set were found to be 0,9760, 1,0000, 0,9906, 0,9823 and 0,9997; respectively. © Springer Science+Business Media, LLC, part of Springer Nature 2020.

Entities:  

Keywords:  Chest X-ray classification; Convolutional neural networks (CNN); Corona 2019; Covid-19; Deep learning; Dual tree complex wavelet transform (DT-CWT); Local binary pattern (LBP)

Year:  2020        PMID: 34764560      PMCID: PMC7609830          DOI: 10.1007/s10489-020-02019-1

Source DB:  PubMed          Journal:  Appl Intell (Dordr)        ISSN: 0924-669X            Impact factor:   5.086


Introduction

In the last few months of 2019, a new type of virus, which is a member of the family Coronaviridae, emerged. The virus in question is considered to have had a zoonotic origin [1]. The virus that emerged in the city of Wuhan in Hubei province in China affected this region first and then spread all over the world in a short time. The virus generally affects the upper and lower respiratory tract, lungs, and, less frequently, the heart muscles [2]. While the virus generally affects young and middle-aged people and people who do not have any chronic diseases to a lesser extent, it can cause severe consequences, resulting in death, in people who suffer from diseases such as hypertension, cardiovascular disease, and diabetes [3]. The epidemic, which was declared to be a pandemic in March 2020 by the World Health Organization; as of the first week of October of the same year, had a number of cases approaching thirty-six million, while the death toll reached one million hundred thousand. Also, a modeling study carried out by Hernandez-Matamoros et al. [4] indicates that the effects of the epidemic will become more severe in the future. In people suffering severely from the disease, the serious adverse effects are generally in the lungs [3]. In this context, many literature studies have been carried out in a short time in which these effects of the disease in the lungs were shown using CT scans of lungs and chest X-ray imaging. Literature studies indicate that radiological imaging, along with clinical symptoms, blood, and biochemical tests, is an effective and reliable diagnostic tool for the diagnosis of Covid-19 disease. Many clinical studies in which X-ray images were examined [5-24] have shown that Covid-19 disease causes interstitial involvement, bilateral and irregular ground-glass opacities, parenchymal abnormalities, a unilobar reversed halo sign, and consolidation on the lungs. The recent review article published by Long and Ehrenfeld [25] highlighted the importance of using artificial intelligence methods to quickly diagnose Covid-19 disease and reduce the effects of the outbreak crisis. In this context, some literature studies have been carried out that diagnose Covid-19 disease (Covid-19 and non-Covid-19) through X-ray images and using deep learning methods. Table 1 contains some summary information about the number of images, study methods, and study results used in these literature studies.
Table 1

Results of previous studies for Covid-19 and non-Covid-19 classification using X-ray images

StudyYearNo. of ImagesMethodsTest MethodsResults
Tuncer et al. [26]2020321 images (87 Covid-19 and 234 Healthy)Residual Exemplar Local Binary Pattern, Iterative Relief, Decision Tree, Linear Discriminant, Support Vector Machine, k-Nearest Neighborhood and Subspace Discriminant10-fold; 80% Train-20% Test; 50% Train-50% TestSen: 0,8755–0,9829/0,8297-0,9798/0,8149-1,0000; Spe: 0,9997-1,0000/0,9444-1,0000/0,9380-1,0000; Acc: 0,9663-0,9955/0,9130-0,9945/0,9049-0,9906
Panwar et al. [27]2020284 images (142 Covid-19 and 142 Healthy)Convolutional Neural Network (nCOVnet)70% Train-30% TestSen: 0,9762; Spe: 0,7857; Acc: 0,881
Ozturk et al. [28]2020625 images (125 Covid-19 and 500 Healthy)Convolutional Neural Network (DarkNet)5-foldSen: 0,9513; Spe: 0,953; Acc: 0,9808; F-1 Score: 0,9651
Mohammed et al. [29]202050 images (25 Covid-19 and 25 Healthy)Multi-Criteria Decision Making (Naive Bayes, Neural Network, Support Vector Machine, Radial Basis Function, k-Nearest Neighbors, Stochastic Gradient Descent, Random Forests, Decision Tree, AdaBoost, CN2 Rule Inducer Algorithm)UnspecifiedSen: 0,706-0,974; Spe: 0,557-1,000; Acc: 0,620-0,987; F-1 Score: 0,555–0,987; AUC: 0,800-0,988; Time: 0,14–7,57 s.
Khan et al. [30]2020594 images (284 Covid-19 and 310 Healthy)Convolutional Neural Network (CoroNet (Xception))4-foldSen: 0,993; Spe: 0,986; Acc: 0,990; F-1 Score: 0,985
Apostolopoulos and Mpesiana [31]2020728 images (224 Covid-19 and 504 Normal)Transfer Learning with Convolutional Neural Networks (VGG19, MobileNet v2, Inception, Xception, Inception ResNet v2)10-foldSen: 0,9866; Spe: 0,9646; Acc: 0,9678
Waheed et al. [32]20201.124 images (403 Covid-19 and 721 Healthy)Convolutional Neural Network (VGG-16) and Synthetic Data AugmentationTrain: 932 (331 Covid-19 and 601 Healthy); Test: 192 (72 Covid-19 and 120 Healthy)Sen: 0,69-0,90; Spe: 0,95-0,97; Acc: 0,85-0,95
Mahmud et al. [33]2020610 images (305 Covid-19 and 305 Healthy)Transfer Learning with Convolutional Neural Networks (Stacked Multi-Resolution CovXNet)5-foldSen: 0,978; Spe: 0,947; Acc: 0,974; F-1 Score: 0,971; AUC: 0,969
Vaid et al. [34]2020545 images (181 Covid-19 and 364 Healthy)Convolutional Neural Network (VGG-19) and Trainable Fully Connected LayersTrain: 348 (115 Covid-19 and 233 Healthy); Validation: 88 (32 Covid-19 and 56 Healthy); Test: 109 (34 Covid-19 and 75 Healthy)Sen: 0,9863; Spe: 0,9166; Acc: 0,9633; F-1 Score: 0,9729
Benbrahim et al. [35]2020320 images (160 Covid-19 and 160 Healthy)Transfer Learning with Convolutional Neural Networks (Inceptionv3 and ResNet50)70% Train-30% TestSen: 0,9803-0,9811; Acc: 0,9803-0,9901; F-1 Score: 0,9803-0,9901
Elaziz et al. [36]2020Dataset-1: 1.891 images (216 Covid-19 and 1.675 Healthy); Dataset-2: 1.560 images (219 Covid-19 and 1.341 Healthy)Fractional Multichannel Exponent Moments, Manta-Ray Foraging Optimization and KNN classifier80% Train-20% TestSen: 0,9875-0,9891; Acc: 0,9609-0,9809
Martínez et al. [37]2020240 images (120 Covid-19 and 120 Healthy)Convolutional Neural Network (Neural Architecture Search Network (NASNet))70% Train-30% TestSen: 0,97; Acc: 0,97; F-1 Score: 0,97
Loey et al. [38]2020148 images (69 Covid-19 and 79 Healthy)Transfer Learning with Convolutional Neural Networks (Alexnet, Googlenet, and Resnet18)Train: 130 (60 Covid-19 and 70 Healthy); Test: 18 (9 Covid-19 and 9 Healthy)Sen: 1,000; Spe:1,000; Acc: 1,000
Toraman et al. [39]20201.281 images (231 Covid-19 and 1.050 Healthy)Convolutional Neural Network (CapsNet)10-foldSen: 0,28-0,9742; Spe:0,8095–0,98; Acc: 0,4914-0,9724; F-1 Score: 0,55-0,9724; Time: 16–500 s. (Note: The results show the average fold.)
Duran-Lopez et al. [40]20206.926 images (2.589 Covid-19 and 4.337 Healthy)Convolutional Neural Network5-foldSen: 0,9253; Spe:0,9633; Acc: 0,9443; F-1 Score: 0,9314; AUC: 0,988
Minaee et al. [41]20205.184 images (184 Covid-19 and 5.000 Healthy)Transfer Learning with Convolutional Neural Networks (ResNet18, ResNet50, SqueezeNet, and DenseNet-121)Train: 2.084 (84 Covid-19 and 2.000 Healthy); Test: 3.100 (100 Covid-19 and 3.000 Healthy)Sen: 0,98; Spe:0,751-0,929
Results of previous studies for Covid-19 and non-Covid-19 classification using X-ray images CT imaging generally contains more data than X-ray imaging. However, it has some disadvantages for the follow-up of all stages of the disease due to the excess amount of radiation that the patients are exposed to. For this reason, an artificial intelligence application using X-ray images was created and tested in the study. In this study, which aims at early diagnosis of Covid-19 disease with the help of X-ray images, a deep learning approach, which is an artificial intelligence method applying the latest technology, was used. In this context, automatic classification of the images was carried out through the two different convolutional neural networks (CNNs). In the study, experiments were carried out for the use of images directly, using local binary pattern (LBP) as a pre-process and dual tree complex wavelet transform (DT-CWT) as a secondary operation, and the results of the automatic classification were calculated separately. Within the scope of the study, four new classification approaches that involve performing the experiments together and combining the results through a result generation algorithm, have been proposed and tested. The results of the study show that in the diagnosis of Covid-19 disease, the analysis of chest X-ray images using deep learning methods provides fast and highly accurate results.

Methods

Used Data

The chest X-ray images of patients with Covid-19 used in the study were obtained by combining metadata data sets that were made open access over GitHub after being created by Cohen et al. [42] and over Kaggle after being created by Dadario [43]. The images that these data sets contain in common and the clinical notes related to these images were combined and a mixed Covid-19 image data set consisting of 150 chest X-ray images was created. In the study, images obtained while the patients were facing the X-ray device directly were used. In the studies, the images taken from the same patient were obtained on different days of the course of the disease and therefore do not contain exactly the same content. The dimensions of the images in question vary between 255 px × 249 px and 4280 px × 3520 px (px is pixel abbreviation) and show a wide variety. Also, these images have different data formats such as png, jpg, jpeg and two different bit depths such as 8-bit (gray-level) and 24-bit (RGB). Standardization of the images is an essential process for use in this study. In this context, all of the images have been converted to 8-bit gray-level images. Then, to clarify the area of interest on the images, manual framing was performed so as to cover the chest area. After this process, all the images were rearranged to 448 px × 448 px and saved in png format. For the non-Covid-19 X-ray images in the study, two data sets, a Montgomery data set [44] and a Shenzhen data set [44], were used separately. These databases contain 80 and 326 non-Covid-19 X-ray images, respectively. The first training-test data set contains a total of 230 X-ray images, of which 150 are Covid-19 images and 80 are non-Covid-19 images, while the second training-test data set contains 476 X-ray images, of which 150 are Covid-19 images and 326 are non-Covid-19 images. Thus, it was ensured that classification results were obtained for the two data sets that contained predominantly Covid-19 images and predominantly non-Covid-19 images, respectively. The processes applied to the Covid-19 images were likewise applied to the non-Covid-19 images. In Fig. 1, original and edited versions of the X-ray images are shown; one belonging to a patient with Covid-19 and two belonging to people without Covid-19 (non-Covid-19 people).
Fig. 1

a) X-ray image of a patient with Covid-19 (Phan et al. [23]) b) Non-Covid-19 X-ray image (Montgomery data set [44])) c) Non-Covid-19 X-ray image (Shenzhen data set [44]))

a) X-ray image of a patient with Covid-19 (Phan et al. [23]) b) Non-Covid-19 X-ray image (Montgomery data set [44])) c) Non-Covid-19 X-ray image (Shenzhen data set [44]))

Local Binary Pattern (LBP)

Local binary pattern (LBP) is an approach that was proposed by Ojala et al. [45] to reveal local features. The method is basically based on comparing a pixel on the image to the neighboring pixels one by one, in terms of size. In Fig. 2, the images obtained by applying the LBP operation to the X-ray images given in Fig. 1 are included. The purpose of benefiting from LBP operation within the scope of this study is to observe the effects of using LBP images, which reflect the local features in the CNN input on the study results, rather than the original images. Additionally, the aim of the study is to increase the image feature depth used in the new result generation algorithm.
Fig. 2

Images created by applying LBP and resizing the images in Fig. 1

Images created by applying LBP and resizing the images in Fig. 1

Dual Tree Complex Wavelet Transform (DT-CWT)

Dual tree complex wavelet transform (DT-CWT) was first introduced by Kingsbury [46-48]. This method is generally similar to the Gabor wavelet transform. In the Gabor wavelet transform, low-pass and high-pass filters are applied to the rows and columns of the image horizontally and vertically. In this way, two different sub-band groups are formed in rows and columns as low (L) and high (H). Crossing is made during the conversion of the said one-dimensional bands into two dimensions. At the end of the process, a low sub-band, named LL, is obtained. In addition, three sub-bands containing high bands, LH, HL, and HH, are formed. Further sub-bands (such as LLL, LLH) can be obtained by applying the same operations to the LL sub-band. Unlike the Gabor wavelet transform, instead of a single filter, DT-CWT uses two filters that work in parallel. These two trees contain real and imaginary parts of complex numbers. That is, as a result of the DT-CWT process, a sub-band containing more directions than the Gabor wavelet transform is obtained. When DT-CWT is applied to an image, the processes are performed for six different directions, +15, −15, +45, −45, +75, and − 75 degrees. Three of these directions represent real sub-bands and the other three represent imaginary sub-bands. Figure 3 shows the DT-CWT decomposition tree. In Fig. 4, real and imaginary sub-band images obtained by applying the DT-CWT process (scale = 1) to the X-ray images given in Fig. 1, are shown. Within the scope of the study, the DT-CWT process was used with a scale (level) value of 1, and the dimensions of the sub-band images obtained were half the size of the original images. Since the complex wavelet transform has been successful in many studies [49-51] where medical images have previously been used, this conversion was preferred in the study.
Fig. 3

Structure of the DT-CWT decomposition tree

Fig. 4

Real and imaginary sub-band images obtained by applying DT-CWT to the X-ray Image (scale = 1)

Structure of the DT-CWT decomposition tree Real and imaginary sub-band images obtained by applying DT-CWT to the X-ray Image (scale = 1)

Convolutional Neural Network (CNN)

Deep learning has come to the fore in recent years as an artificial intelligence approach that provides successful results in many image processing applications from image enhancement (such as [52]) to object identification (such as [53, 54]). Convolutional neural network (CNN) has been the preferred deep learning model in image processing applications in recent years. The CNN classifier, in general, consists of a convolution layer, activation functions, a pooling layer, a flatten layer, and fully connected layer components. In this context, Fig. 5 describes the general operation of the CNN classifier. It is possible to examine more detailed information about the functions and operating modes of the layers in the CNN classifier from the studies [55-59].
Fig. 5

General operation of the CNN classifier

General operation of the CNN classifier Within the scope of the study, a CNN architecture with a total of 23 layers was designed. An effective design was aimed at, since increasing the number of layers in the CNN architecture leads to increased processing time in the training and classification processes. Table 2 contains details of the first CNN architecture used in the study. Also, a second CNN architecture was used to check whether the proposed pipeline approaches applied to other CNN architectures. In this context, an architecture modeled on VGG-16 CNN was used. However, to reduce the processing load, the number of filters and the fully connected layer sizes have been reduced. Additionally, normalization layers were added after the intermediate convolution layers. Details of this second CNN architecture used are given in Table 3.
Table 2

First CNN architecture used within the scope of the study

LayerLayer NameLayer Parameters (Matlab)
1imageInputLayer[448,448 1], [224,224 1], [224,224 2], [224,224 3] and [224,224 6]
2convolution2dLayer(3,4,'Padding’,'same’)
3batchNormalizationLayerdefault
4reluLayerdefault
5maxPooling2dLayer(2,'Stride’,2)
6convolution2dLayer(3,8,'Padding’,'same’)
7batchNormalizationLayerdefault
8reluLayerdefault
9maxPooling2dLayer(2,'Stride’,2)
10convolution2dLayer(3,16,'Padding’,'same’)
11batchNormalizationLayerdefault
12reluLayerdefault
13maxPooling2dLayer(2,'Stride’,2)
14convolution2dLayer(3,32,'Padding’,'same’)
15batchNormalizationLayerdefault
16reluLayerdefault
17maxPooling2dLayer(2,'Stride’,2)
18convolution2dLayer(3,64,'Padding’,'same’)
19batchNormalizationLayerdefault
20reluLayerdefault
21fullyConnectedLayer2
22softmaxLayerdefault
23classificationLayerdefault
Table 3

Second CNN architecture used within the scope of the study

LayerLayer NameLayer Parameters (Matlab)
1imageInputLayer[448,448 1], [224,224 1], [224,224 2], [224,224 3] and [224,224 6]
2convolution2dLayer(3,8,'Padding’,'same’)
3batchNormalizationLayerdefault
4reluLayerdefault
5convolution2dLayer(3,8,'Padding’,'same’)
6batchNormalizationLayerdefault
7reluLayerdefault
8maxPooling2dLayer(2,'Stride’,2)
9convolution2dLayer(3,16,'Padding’,'same’)
10batchNormalizationLayerdefault
11reluLayerdefault
12convolution2dLayer(3,16,'Padding’,'same’)
13batchNormalizationLayerdefault
14reluLayerdefault
15maxPooling2dLayer(2,'Stride’,2)
16convolution2dLayer(3,32,'Padding’,'same’)
17batchNormalizationLayerdefault
18reluLayerdefault
19convolution2dLayer(3,32,'Padding’,'same’)
20batchNormalizationLayerdefault
21reluLayerdefault
22convolution2dLayer(3,32,'Padding’,'same’)
23batchNormalizationLayerdefault
24reluLayerdefault
25maxPooling2dLayer(2,'Stride’,2)
26convolution2dLayer(3,64,'Padding’,'same’)
27batchNormalizationLayerdefault
28reluLayerdefault
29convolution2dLayer(3,64,'Padding’,'same’)
30batchNormalizationLayerdefault
31reluLayerdefault
32convolution2dLayer(3,64,'Padding’,'same’)
33batchNormalizationLayerdefault
34reluLayerdefault
35maxPooling2dLayer(2,'Stride’,2)
36convolution2dLayer(3,64,'Padding’,'same’)
37batchNormalizationLayerdefault
38reluLayerdefault
39convolution2dLayer(3,64,'Padding’,'same’)
40batchNormalizationLayerdefault
41reluLayerdefault
42convolution2dLayer(3,64,'Padding’,'same’)
43batchNormalizationLayerdefault
44reluLayerdefault
45maxPooling2dLayer(2,'Stride’,2)
46fullyConnectedLayer512
47reluLayerdefault
48dropoutLayer0,5
49fullyConnectedLayer512
50reluLayerdefault
51dropoutLayer0,5
52fullyConnectedLayer2
53softmaxLayerdefault
54classificationLayerdefault
First CNN architecture used within the scope of the study Second CNN architecture used within the scope of the study In the context of the study, Matlab 2019a program was preferred as software. The layer names and parameters in Tables 2 and 3 are the names and parameters used directly in the software. In the study, more than one experiment was carried out and the sizes of the input images used in the experiments differ. For this reason, there are different sizes in the input layer in Tables 2 and 3. Those CNN architectures were used in all the experiments carried out within the scope of the study.

Evaluation Criteria of the Classification Results

Within the scope of the study, confusion matrix and statistical parameters obtained from this matrix were used to evaluate the results. It is possible to examine detailed information about the confusion matrix, i.e., sensitivity (SEN), specificity (SPE), accuracy (ACC), and F-1 score (F-1), from the studies [60]. Receiver operating characteristic (ROC) analysis was also used to evaluate the results. In addition, the sizes of the areas under the ROC curve (Area Under Curve (AUC)) were calculated. ROC analysis basically reflects graphically the variation of sensitivity (SEN) (y-axis) relative to 1-SPE (x-axis) for the case that the threshold value is gradually changed with a certain precision between the minimum and maximum output predicted for the classification.

Pipeline Methodology

First of all, in the proposed pipeline algorithm, training and test procedures for images of size of 448 × 448 were performed and results were obtained. Before the experiments after the first experiment were conducted, DT-CWT was applied to the images of size of 448 px × 448 px (scale = 1) and 224 px × 224 px sub-band images were obtained. In the second experiment, results were obtained for the case of giving the real part of the LL sub-band image obtained by applying DT-CWT, as input to the CNN. In the third experiment, training and testing procedures were carried out and results were obtained for the case of giving the imaginary part of the LL sub-band image obtained by applying DT-CWT, as input to the CNN. In the fourth experiment, training and testing procedures were carried out and results were obtained for the case of giving the real parts of LL, LH and HL sub-band images obtained by applying DT-CWT as input to the CNN, together. In the fifth experiment, training and test procedures were carried out and results were obtained for the case of giving the imaginary parts of the LL, LH and HL sub-band images obtained by applying DT-CWT as input to the CNN, together. In the sixth experiment, results were obtained for the case of giving the real and imaginary parts of the LL sub-band image obtained by applying DT-CWT as input to the CNN, together. In the seventh experiment, results were obtained for the case of giving the real and imaginary parts of the LL, LH, HL sub-band images obtained by applying DT-CWT, as input to the CNN, together. A block diagram of the experiments carried out in the study is shown in Fig. 6. The first seven experiments conducted were repeated using new images obtained by applying LBP to the X-ray images, and the first stage experiments were completed. Since the image size decreases after LBP processing, these images were rearranged as 448 px × 448 px in size.
Fig. 6

Block diagram representation of the study of the experiments

Block diagram representation of the study of the experiments In the ongoing part of the study, four pipeline classification algorithms were designed using the principle of parallel operation. These algorithms are based on combining the results of previous experiments to obtain new results. The first two pipeline classification algorithms mentioned above work as follows: If the numbers of labeling (threshold value for 0,5) obtained in the experiments (with and without LBP) for an image are not equal to each other, the labeling result obtained in more than half of the experiments is considered to be the algorithm labeling result for Covid-19 or non-Covid-19. In this case, if the algorithm labeling result is Covid-19, the real experiment result is the closest to the number 0; if the algorithm labeling result is non-Covid-19, the closest to the number 1, respectively, is assigned as the algorithm result. If the numbers of labeling (for threshold value 0,5) obtained in the experiments (with and without LBP) for an image are equal to each other, the actual test results obtained in the experiments conducted for the image are mixed (50%–50% and 75%–25%, respectively), and the result is accepted as the algorithm result. After that, the labeling of the image is realized as Covid-19 or non-Covid-19 (for the threshold value 0,5) according to this result obtained. The basic coding of the first two pipeline classification approaches is included in Table 4. In the codes between Tables 4 and 6, Result-1 and Label-1 represent the actual test result and the label obtained without using LBP, while Result-2 and Label-2 represent the actual test result and the label obtained using LBP.
Table 4

Basic coding of the pipeline algorithms (pipeline-1 and -2) proposed in the study

Table 6

Basic coding of the pipeline algorithm (pipeline-4) proposed in the study

Basic coding of the pipeline algorithms (pipeline-1 and -2) proposed in the study In the third and fourth pipeline algorithms, unlike the first two pipeline algorithms, if the tags obtained as a result of the classification experiment differ from each other, the result obtained without applying LBP has been taken into consideration with priority. Accordingly, in the case where the two classification tags are different from each other in the third pipeline algorithm, if the tag result obtained without applying LBP was abnormal, the result was considered abnormal. In the fourth pipeline algorithm, in the case of the two classification tags being different from each other, if the tag result obtained without applying LBP was normal, the result was considered normal. The other procedures are the same as for the first two pipeline algorithms. A mixing rate of 50% -50% was applied in the third and fourth pipeline algorithms. The basic coding of the third and fourth pipeline classification approaches is given in Tables 5 and 6.
Table 5

Basic coding of the pipeline algorithm (pipeline-3) proposed in the study

Basic coding of the pipeline algorithm (pipeline-3) proposed in the study Basic coding of the pipeline algorithm (pipeline-4) proposed in the study

Results

Experiments

In this study, which aims to detect Covid-19 disease early using X-ray images, the deep learning approach, which is the artificial intelligence method applying the latest technology, was used and automatic classification of the images was performed using CNN. In the first training-test data set used in the study, there were 230 X-ray images, of which 150 were Covid-19 and 80 were non-Covid-19, while in the second training-test data set there were 476 X-ray images, of which 150 were Covid-19 and 326 were non-Covid-19. Thus, it was ensured that the classification results were obtained separately from the two data sets containing predominantly abnormal images and predominantly normal images. The information from the training-test data sets is given in Table 7.
Table 7

Information about the images used in the study

Data setCovid-19 ImageNormal ImageTotal
1. Data set150 ([42, 43])80 (Montgomery [44])230
2. Data set150 ([42, 43])326 (Shenzhen [44])476
Information about the images used in the study Within the scope of the study, chest X-ray images were manually framed to cover the lung region, primarily to determine the areas of interest on the image. Then, standardization was carried out since the images used were of very different sizes, formats, and bit depths. The areas of interest on the image were resized and the image sizes were arranged as 448 px × 448 px. After that, the images in question were saved in png format so as to be as gray-scale and 8-bit depth. These operations were applied to all the abnormal and normal images used in the study. In the ongoing part of the study, a 23-layer CNN architecture and a 54-layer CNN architecture were designed and used, the details of which have been previously described. Those CNN architectures were used in all the experiments. Due to the fact that more than one experiment was performed within the scope of the study, only the images given to the CNN input differ in size. In the experiments conducted in the study, the trainings were carried out with the k-fold cross validation method. In this context, the k value was chosen as 23. Since the first training-test data set consists of 230 images, 220 images, except for ten images at each stage (fold), were used for the training operations, and the remaining ten images were used for the testing operations. The second training-test data set consists of 476 images, and, in the same way, except 20/21 (16 groups consisting of 21 images and seven groups of 20 images) images, 456/455 images were used in the training operations, and the remaining 20/21 images were used in the testing operations. The test procedures were repeated 23 times and classification results were obtained for all the images. Finally, within the scope of the study, all the images were combined and the training and testing procedures were repeated by applying a 2-fold cross for a total of 556 X-ray images comprising 150 Covid-19 images and 406 non-Covid-19 images. Considering the length of the study as well, the results that have been shared in the study are only for the input data that provided the best results for the first and second data sets. In this part of the study, a total of 14 experiments were carried out. Some initial weights and parameters in the CNN are randomly assigned. To make the study results stable, each experiment was repeated five times in itself, and average results in the study are shown. Within the scope of the study, the CPU time taken for an experiment to be completed entirely, including the training and testing, was divided by the total number of images processed, and the processing CPU time per image was measured. The experiments of this study were carried out using MATLAB 2019 (a) software running on a computer with 64 GB RAM and Intel(R) Xeon (R) CPU E5–2680 2.7 GHz (32 CPUs). In the first experimental group within the scope of the study, the training and testing procedures were first performed using the chest X-ray images, and the results were obtained. LBP operation was then applied to the images in question, and then the training and testing procedures were repeated and the results were calculated. Finally, the results were calculated using the pipeline classification algorithms, the details of which were previously described and proposed within the scope of the study. Due to the random assignment of some initial variables used in the internal structure of the CNN, each experiment group was repeated five times in order to make the results more stable. The image sizes given to the CNN as input for this experiment were 448 × 448 × 1. The results obtained from the experimental group are given in Table 8 (first training-test data set) and Table 9 (second training-test data set).
Table 8

Results obtained directly using chest X-ray images (first training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP146,23,876,83,20,97470,96000,96960,97660,996118,964
With LBP147,82,271,68,40,98530,89500,95390,96550,993918,226
Pipeline-1148,61,476,23,80,99070,95250,97740,98280,996837,190
Pipeline-2146,83,276,63,40,97870,95750,97130,97800,996837,190
Pipeline-3148,81,275,84,20,99200,94750,97650,98220,996837,190
Pipeline-4146,04,077,22,80,97330,96500,97040,97720,996937,190
Second CNN ArchitectureWithout LBP144,25,876,63,40,96130,95750,96000,96910,994933,105
With LBP142,67,474,06,00,95070,92500,94170,95510,981536,433
Pipeline-1147,22,878,41,60,98130,98000,98090,98530,997769,538
Pipeline-2144,45,677,42,60,96270,96750,96430,97240,997369,538
Pipeline-3148,02,076,63,40,98670,95750,97650,98210,997269,538
Pipeline-4143,46,678,41,60,95600,98000,96430,97220,995869,538
Table 9

Results obtained directly using chest X-ray images (second training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP144,45,6322,04,00,96270,98770,97980,96780,997524,878
With LBP137,212,8316,49,60,91470,97060,95290,92460,989924,995
Pipeline-1146,04,0322,04,00,97330,98770,98320,97330,998249,872
Pipeline-2146,04,0321,84,20,97330,98710,98280,97260,998349,872
Pipeline-3147,03,0321,05,00,98000,98470,98320,97350,998149,872
Pipeline-4143,46,6323,03,00,95600,99080,97980,96760,997949,872
Second CNN ArchitectureWithout LBP146,04,0323,03,00,97330,99080,98530,97660,997751,817
With LBP133,816,2315,610,40,89200,96810,94410,90960,988554,239
Pipeline-1148,02,0322,63,40,98670,98960,98870,98210,998710,6056
Pipeline-2147,03,0323,22,80,98000,99140,98780,98060,998810,6056
Pipeline-3148,81,2322,04,00,99200,98770,98910,98280,999110,6056
Pipeline-4145,24,8323,62,40,96800,99260,98490,97580,997510,6056
Results obtained directly using chest X-ray images (first training-test data set) Results obtained directly using chest X-ray images (second training-test data set) In the second experimental group within the scope of the study, the training and testing procedures were performed using the real part of the LL sub-image obtained by applying DT-CWT to the chest X-ray images, and the results were obtained. Then, the training and testing procedures were performed using the real part of the LL sub-image obtained by applying the LBP and DT-CWT operations to the X-ray images, respectively. Finally, the results were calculated using the pipeline classification algorithms, the details of which were previously described and proposed within the scope of the study. The image sizes given to the CNN as input for this experiment were 224 × 224 × 1. The results obtained from the experimental group are given in Table 10 (first training-test data set) and Table 11 (second training-test data set).
Table 10

Results obtained by using the LL real sub-band obtained by applying DT-CWT to the chest X-ray images (first training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP146,63,477,42,60,97730,96750,97390,98000,99830,6090
With LBP146,04,072,67,40,97330,90750,95040,96240,98900,6252
Pipeline-1149,01,077,42,60,99330,96750,98430,98810,998812,342
Pipeline-2147,03,077,42,60,98000,96750,97570,98130,998912,342
Pipeline-3149,20,876,63,40,99470,95750,98170,98620,999012,342
Pipeline-4146,43,678,21,80,97600,97750,97650,98190,998612,342
Second CNN ArchitectureWithout LBP141,88,277,82,20,94530,97250,95480,96460,993412,146
With LBP141,28,863,616,40,94130,79500,89040,91830,958012,339
Pipeline-1146,63,477,62,40,97730,97000,97480,98060,996324,485
Pipeline-2144,06,078,02,00,96000,97500,96520,97290,996224,485
Pipeline-3147,42,677,22,80,98270,96500,97650,98200,996124,485
Pipeline-4141,09,078,21,80,94000,97750,95300,96310,994424,485
Table 11

Results obtained by using the LL real sub-band obtained by applying DT-CWT to the chest X-ray images (second training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP143,66,4320,85,20,95730,98400,97560,96110,99730,7571
With LBP135,214,8311,214,80,90130,95460,93780,90130,98650,7569
Pipeline-1146,43,6322,23,80,97600,98830,98450,97530,997815,140
Pipeline-2145,05,0321,64,40,96670,98650,98030,96850,997915,140
Pipeline-3146,83,2320,45,60,97870,98280,98150,97090,998115,140
Pipeline-4143,26,8322,63,40,95470,98960,97860,96560,997315,140
Second CNN ArchitectureWithout LBP146,04,0322,04,00,97330,98770,98320,97330,998315,029
With LBP131,818,2305,420,60,87870,93680,91850,87180,977018,120
Pipeline-1147,62,4321,64,40,98400,98650,98570,97750,997733,149
Pipeline-2146,63,4322,43,60,97730,98900,98530,97670,998033,149
Pipeline-3148,21,8320,85,20,98800,98400,98530,97700,998433,149
Pipeline-4145,44,6322,83,20,96930,99020,98360,97390,997533,149
Results obtained by using the LL real sub-band obtained by applying DT-CWT to the chest X-ray images (first training-test data set) Results obtained by using the LL real sub-band obtained by applying DT-CWT to the chest X-ray images (second training-test data set) In the third experimental group within the scope of the study, the training and testing procedures were performed using the imaginary part of the LL sub-image obtained by applying DT-CWT to the chest X-ray images, and the results were obtained. Then, the training and testing procedures were performed using the imaginary part of the LL sub-image obtained by applying the LBP and DT-CWT operations to the X-ray images, respectively. Finally, the results were calculated using the pipeline classification algorithms, the details of which were previously described and proposed within the scope of the study. The image sizes given to the CNN as input for this experiment were 224 × 224 × 1. The results obtained from the experimental group are given in Table 12 (first training-test data set) and Table 13 (second training-test data set).
Table 12

Results obtained by using the LL imaginary sub-band obtained by applying DT-CWT to the chest X-ray images (first training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP146,43,677,42,60,97600,96750,97300,97920,99720,6033
With LBP145,64,473,66,40,97070,92000,95300,96430,98890,5979
Pipeline-1147,82,277,03,00,98530,96250,97740,98270,998012,012
Pipeline-2146,63,477,42,60,97730,96750,97390,97990,998112,012
Pipeline-3148,21,876,43,60,98800,95500,97650,98210,998512,012
Pipeline-4146,04,078,02,00,97330,97500,97390,97990,997712,012
Second CNN ArchitectureWithout LBP141,09,077,22,80,94000,96500,94870,95990,980312,156
With LBP143,66,460,219,80,95730,75250,88610,91670,964512,086
Pipeline-1146,83,275,84,20,97870,94750,96780,97540,992324,243
Pipeline-2143,07,077,42,60,95330,96750,95830,96750,990924,243
Pipeline-3146,83,274,85,20,97870,93500,96350,97220,989824,243
Pipeline-4141,09,078,21,80,94000,97750,95300,96310,984924,243
Table 13

Results obtained by using the LL imaginary sub-band obtained by applying DT-CWT to the chest X-ray images (second training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP145,05,0322,04,00,96670,98770,98110,96990,99700,7587
With LBP134,815,2312,014,00,89870,95710,93870,90220,98610,7578
Pipeline-1147,22,8322,04,00,98130,98770,98570,97740,998115,164
Pipeline-2146,04,0322,23,80,97330,98830,98360,97400,998215,164
Pipeline-3147,62,4321,05,00,98400,98470,98450,97550,998515,164
Pipeline-4144,65,4323,03,00,96400,99080,98240,97180,996915,164
Second CNN ArchitectureWithout LBP146,04,0323,22,80,97330,99140,98570,97720,998115,022
With LBP128,821,2306,020,00,85870,93870,91340,86150,977316,851
Pipeline-1147,42,6322,63,40,98270,98960,98740,98000,997931,874
Pipeline-2146,23,8323,03,00,97470,99080,98570,97730,998431,874
Pipeline-3147,82,2322,04,00,98530,98770,98700,97940,998331,874
Pipeline-4145,64,4323,82,20,97070,99330,98610,97780,998231,874
Results obtained by using the LL imaginary sub-band obtained by applying DT-CWT to the chest X-ray images (first training-test data set) Results obtained by using the LL imaginary sub-band obtained by applying DT-CWT to the chest X-ray images (second training-test data set) In the fourth experimental group within the scope of the study, the training and testing procedures were performed using the real part of the LL, LH and HL sub-images obtained by applying DT-CWT to the chest X-ray images, and the results were obtained. Then, the training and testing procedures were performed using the real part of the LL, LH and HL sub-images obtained by applying the LBP and DT-CWT operations to the X-ray images, respectively. Finally, the results were calculated using the pipeline classification algorithms, the details of which were previously described and proposed within the scope of the study. The image sizes given to the CNN as input for this experiment were 224 × 224 × 3. The results obtained from the experimental group are given in Table 14 (first training-test data set) and Table 15 (second training-test data set).
Table 14

Results obtained using the LL, LH, HL real sub-bands obtained by applying DT-CWT to the chest X-ray images (first training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP146,23,877,82,20,97470,97250,97390,97990,997610,596
With LBP147,42,671,68,40,98270,89500,95220,96420,992610,598
Pipeline-1148,41,677,42,60,98930,96750,98170,98600,998421,193
Pipeline-2147,03,077,82,20,98000,97250,97740,98260,998421,193
Pipeline-3148,41,677,03,00,98930,96250,98000,98470,998321,193
Pipeline-4146,23,878,21,80,97470,97750,97570,98120,998321,193
Second CNN ArchitectureWithout LBP142,67,476,83,20,95070,96000,95390,96420,987616,868
With LBP141,28,864,215,80,94130,80250,89300,92010,966416,890
Pipeline-1146,83,276,43,60,97870,95500,97040,97740,994833,757
Pipeline-2143,86,276,83,20,95870,96000,95910,96840,994433,757
Pipeline-3147,42,675,84,20,98270,94750,97040,97750,994533,757
Pipeline-4142,08,077,42,60,94670,96750,95390,96400,990333,757
Table 15

Results obtained using the LL, LH, HL real sub-bands obtained by applying DT-CWT to the chest X-ray images (second training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP145,44,6322,83,20,96930,99020,98360,97390,997413,986
With LBP132,018,0309,816,20,88000,95030,92820,88520,980714,043
Pipeline-1146,63,4320,85,20,97730,98400,98190,97150,997928,029
Pipeline-2146,43,6322,63,40,97600,98960,98530,97660,998228,029
Pipeline-3147,82,2320,65,40,98530,98340,98400,97490,998628,029
Pipeline-4144,25,8323,03,00,96130,99080,98150,97040,997028,029
Second CNN ArchitectureWithout LBP144,65,4323,42,60,96400,99200,98320,97300,998521,551
With LBP135,015,0308,217,80,90000,94540,93110,89180,982821,720
Pipeline-1146,63,4323,03,00,97730,99080,98660,97860,998743,271
Pipeline-2145,64,4323,03,00,97070,99080,98450,97520,998943,271
Pipeline-3147,22,8322,43,60,98130,98900,98660,97870,998743,271
Pipeline-4144,06,0324,02,00,96000,99390,98320,97290,998743,271
Results obtained using the LL, LH, HL real sub-bands obtained by applying DT-CWT to the chest X-ray images (first training-test data set) Results obtained using the LL, LH, HL real sub-bands obtained by applying DT-CWT to the chest X-ray images (second training-test data set) In the fifth experimental group within the scope of the study, the training and testing procedures were performed using the imaginary part of the LL, LH and HL sub-images obtained by applying DT-CWT to the chest X-ray images, and the results were obtained. Then, the training and testing procedures were performed using the imaginary part of the LL, LH and HL sub-images obtained by applying the LBP and DT-CWT operations to the X-ray images, respectively. Finally, the results were calculated using the pipeline classification algorithms, the details of which were previously described and proposed within the scope of the study. The image sizes given to the CNN as input for this experiment were 224 × 224 × 3. The results obtained from the experimental group are given in Table 16 (first training-test data set) and Table 17 (second training-test data set).
Table 16

Results obtained using the LL, LH, HL imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (first training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP145,24,877,42,60,96800,96750,96780,97510,996710,667
With LBP146,83,271,88,20,97870,89750,95040,96280,991610,593
Pipeline-1148,41,677,42,60,98930,96750,98170,98610,998221,261
Pipeline-2145,84,277,42,60,97200,96750,97040,97720,998021,261
Pipeline-3148,61,476,83,20,99070,96000,98000,98480,998321,261
Pipeline-4145,05,078,02,00,96670,97500,96960,97640,997021,261
Second CNN ArchitectureWithout LBP143,66,476,23,80,95730,95250,95570,96570,987616,952
With LBP140,49,666,813,20,93600,83500,90090,92520,968416,866
Pipeline-1146,43,676,83,20,97600,96000,97040,97730,994033,818
Pipeline-2144,25,876,04,00,96130,95000,95740,96710,994333,818
Pipeline-3147,62,475,05,00,98400,93750,96780,97550,993633,818
Pipeline-4142,47,678,02,00,94930,97500,95830,96740,990733,818
Table 17

Results obtained using the LL, LH, HL imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (second training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP145,84,2322,83,20,97200,99020,98450,97520,998014,026
With LBP134,815,2311,414,60,89870,95520,93740,90040,983914,067
Pipeline-1146,23,8322,04,00,97470,98770,98360,97400,998242,160
Pipeline-2146,04,0322,43,60,97330,98900,98400,97460,998370,253
Pipeline-3147,42,6321,24,80,98270,98530,98450,97550,998811,2414
Pipeline-4144,65,4323,62,40,96400,99260,98360,97370,997718,2667
Second CNN ArchitectureWithout LBP145,24,8323,03,00,96800,99080,98360,97380,998721,541
With LBP130,020,0304,421,60,86670,93370,91260,86150,973321,514
Pipeline-1147,22,8321,84,20,98130,98710,98530,97680,997943,054
Pipeline-2145,44,6322,83,20,96930,99020,98360,97390,998443,054
Pipeline-3147,82,2321,64,40,98530,98650,98610,97820,998643,054
Pipeline-4144,65,4323,22,80,96400,99140,98280,97240,998243,054
Results obtained using the LL, LH, HL imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (first training-test data set) Results obtained using the LL, LH, HL imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (second training-test data set) In the sixth experimental group within the scope of the study, the training and testing procedures were performed using the real and imaginary parts of the LL sub-image obtained by applying DT-CWT to the chest X-ray images, and the results were obtained. Then, the training and testing procedures were performed using the real and imaginary parts of the LL sub-image obtained by applying the LBP and DT-CWT operations to the X-ray images, respectively. Finally, the results were calculated using the pipeline classification algorithms, the details of which were previously described and proposed within the scope of the study. The image sizes given to the CNN as input for this experiment were 224 × 224 × 2. The results obtained from the experimental group are given in Table 18 (first training-test data set) and Table 19 (second training-test data set).
Table 18

Results obtained by using the LL real and imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (first training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP145,84,277,62,40,97200,97000,97130,97790,99700,8271
With LBP145,84,268,811,20,97200,86000,93300,95000,98830,8273
Pipeline-1148,21,877,22,80,98800,96500,98000,98470,998216,544
Pipeline-2147,03,078,02,00,98000,97500,97830,98330,998316,544
Pipeline-3148,41,676,43,60,98930,95500,97740,98280,998616,544
Pipeline-4145,64,478,41,60,97070,98000,97390,97980,997016,544
Second CNN ArchitectureWithout LBP143,07,076,43,60,95330,95500,95390,96430,990214,416
With LBP142,67,467,212,80,95070,84000,91220,93400,973814,399
Pipeline-1148,02,077,42,60,98670,96750,98000,98470,996928,815
Pipeline-2144,25,877,42,60,96130,96750,96350,97170,995728,815
Pipeline-3148,02,075,64,40,98670,94500,97220,97890,995728,815
Pipeline-4143,07,078,21,80,95330,97750,96170,97010,992228,815
Table 19

Results obtained by using the LL real and imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (second training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP143,07,0323,03,00,95330,99080,97900,96620,998010,808
With LBP135,214,8313,612,40,90130,96200,94290,90860,986310,820
Pipeline-1146,83,2321,84,20,97870,98710,98450,97540,998321,627
Pipeline-2145,44,6323,03,00,96930,99080,98400,97450,998321,627
Pipeline-3147,22,8321,44,60,98130,98590,98450,97550,998721,627
Pipeline-4142,67,4323,42,60,95070,99200,97900,96610,997721,627
Second CNN ArchitectureWithout LBP145,24,8322,83,20,96800,99020,98320,97320,998218,172
With LBP131,218,8306,020,00,87470,93870,91850,87110,979718,135
Pipeline-1147,42,6323,03,00,98270,99080,98820,98130,998536,307
Pipeline-2145,64,4323,22,80,97070,99140,98490,97580,998636,307
Pipeline-3148,02,0321,84,20,98670,98710,98700,97950,999036,307
Pipeline-4144,65,4324,02,00,96400,99390,98450,97500,998036,307
Results obtained by using the LL real and imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (first training-test data set) Results obtained by using the LL real and imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (second training-test data set) In the seventh experimental group within the scope of the study, the training and testing procedures were performed using the real and imaginary parts of the LL, LH, HL sub-images obtained by applying DT-CWT to the chest X-ray images, and the results were obtained. Then, the training and testing procedures were performed using the real and imaginary parts of the LL, LH, HL sub-images obtained by applying the LBP and DT-CWT operations to the X-ray images, respectively. Finally, the results were calculated using the pipeline classification algorithms, the details of which were previously described and proposed within the scope of the study. The image sizes given to the CNN as input for this experiment were 224 × 224 × 6. The results obtained from he experimental group are given in Table 20 (first training-test data set) and Table 21 (second training-test data set).
Table 20

Results obtained by using the LL, LH, HL real and imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (first training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP147,03,077,62,40,98000,97000,97650,98190,997517,203
With LBP145,44,674,25,80,96930,92750,95480,96550,990217,311
Pipeline-1148,21,877,42,60,98800,96750,98090,98540,998434,513
Pipeline-2147,22,877,82,20,98130,97250,97830,98330,998434,513
Pipeline-3148,81,277,22,80,99200,96500,98260,98670,998634,513
Pipeline-4146,43,677,82,20,97600,97250,97480,98060,998134,513
Second CNN ArchitectureWithout LBP142,87,276,83,20,95200,96000,95480,96480,988424,206
With LBP142,08,070,49,60,94670,88000,92350,94170,976324,210
Pipeline-1146,63,477,62,40,97730,97000,97480,98060,995648,417
Pipeline-2144,65,477,42,60,96400,96750,96520,97310,994848,417
Pipeline-3147,82,276,43,60,98530,95500,97480,98080,995348,417
Pipeline-4141,68,478,02,00,94400,97500,95480,96460,990948,417
Table 21

Results obtained by using the LL, LH, HL real and imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (second training-test data set)

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP144,85,2323,03,00,96530,99080,98280,97240,998223,185
With LBP129,220,8308,617,40,86130,94660,91970,87100,978623,401
Pipeline-1147,03,0321,64,40,98000,98650,98450,97550,997446,586
Pipeline-2145,24,8322,83,20,96800,99020,98320,97320,997846,586
Pipeline-3147,62,4321,44,60,98400,98590,98530,97680,998446,586
Pipeline-4144,25,8323,22,80,96130,99140,98190,97100,997546,586
Second CNN ArchitectureWithout LBP144,65,4323,42,60,96400,99200,98320,97310,998431,573
With LBP131,418,6304,821,20,87600,93500,91640,86860,980631,369
Pipeline-1147,42,6322,43,60,98270,98900,98700,97940,998262,942
Pipeline-2145,05,0322,63,40,96670,98960,98240,97190,998662,942
Pipeline-3148,21,8321,84,20,98800,98710,98740,98020,998662,942
Pipeline-4143,86,2324,02,00,95870,99390,98280,97230,998362,942
Results obtained by using the LL, LH, HL real and imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (first training-test data set) Results obtained by using the LL, LH, HL real and imaginary sub-bands obtained by applying DT-CWT to the chest X-ray images (second training-test data set) Finally, all the training-test data sets were combined to test the performance of the proposed method and the pipeline approaches. In this context, a collective training-test data set containing a total of 556 X-ray images comprising 150 Covid-19 and 406 non-Covid-19 images was created. Then the k value was determined as 2 (cross training and testing for 75 Covid-19 and 203 non-Covid-19 images). The training and testing processes were realized for the input images (original image and the LL (real sub-band)), ensuring the best results in the first and second training-test data sets. The results obtained are given in Tables 22 and 23.
Table 22

Results obtained directly using chest X-ray images (k = 2 and a total of 556 images (150 Covid-19 and 406 non-Covid-19 images))

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP138,811,2401,64,40,92530,98920,97190,94680,99370,1460
With LBP133,416,6400,85,20,88930,98720,96080,92410,99320,1401
Pipeline-1144,85,2406,00,00,96531,00000,99060,98230,99970,2861
Pipeline-2139,810,2402,43,60,93200,99110,97520,95290,99910,2861
Pipeline-3146,43,6401,64,40,97600,98920,98560,97340,99840,2861
Pipeline-4137,212,8406,00,00,91471,00000,97700,95540,99770,2861
Second CNN ArchitectureWithout LBP138,611,4403,22,80,92400,99310,97450,95110,99490,2853
With LBP120,629,4395,011,00,80400,97290,92730,85650,97860,2811
Pipeline-1143,07,0405,40,60,95330,99850,98630,97400,99940,5664
Pipeline-2141,09,0404,61,40,94000,99660,98130,96440,99920,5664
Pipeline-3145,64,4403,03,00,97070,99260,98670,97520,99910,5664
Pipeline-4136,014,0405,60,40,90670,99900,97410,94950,99640,5664
Table 23

Results obtained using the LL real sub-band obtained by applying DT-CWT to the chest X-ray images (k = 2 and a total of 556 images (150 Covid-19 and 406 non-Covid-19 images))

CNN TypeMethodTPFNTNFPSENSPEACCF-1AUCCPU Time
First CNN ArchitectureWithout LBP136,213,8401,44,60,90800,98870,96690,93660,99390,0496
With LBP121,628,4396,010,00,81070,97540,93090,86340,97980,0492
Pipeline-1140,89,2405,40,60,93870,99850,98240,96630,99870,0987
Pipeline-2138,811,2403,22,80,92530,99310,97480,95190,99850,0987
Pipeline-3142,87,2401,44,60,95200,98870,97880,96030,99810,0987
Pipeline-4134,215,8405,40,60,89470,99850,97050,94230,99650,0987
Second CNN ArchitectureWithout LBP137,812,2403,42,60,91870,99360,97340,94860,99750,0932
With LBP99,250,8395,011,00,66130,97290,88880,75800,95530,0920
Pipeline-1136,413,6406,00,00,90931,00000,97550,95230,99880,1852
Pipeline-2139,210,8404,61,40,92800,99660,97810,95770,99870,1852
Pipeline-3141,09,0403,42,60,94000,99360,97910,96030,99830,1852
Pipeline-4133,216,8406,00,00,88801,00000,96980,94010,99870,1852
Results obtained directly using chest X-ray images (k = 2 and a total of 556 images (150 Covid-19 and 406 non-Covid-19 images)) Results obtained using the LL real sub-band obtained by applying DT-CWT to the chest X-ray images (k = 2 and a total of 556 images (150 Covid-19 and 406 non-Covid-19 images))

Conclusion

In this section, first of all, the results that were obtained without using pipeline algorithms are compared. When the results of the study given between Tables 8 and 23 are examined within the scope of the study, it can be seen that the results of the study obtained without using LBP are generally better than the results of the study using LBP, for the same input image. In this context, it is understood that there are exceptions for the sensitivity parameter of some results obtained using the first CNN architecture for the first training-test data set. Within the scope of the study, the highest mean sensitivity, specificity, accuracy, F-1 score, and AUC values obtained without using the pipeline algorithms were, respectively; 0,9853, 0,9725, 0,9765, 0,9819, 0,9983 for the first training-test data set and the first CNN architecture, 0,9613, 0,9725, 0,9600, 0,9691, 0,9949 for the first training-test data set and the second CNN architecture, 0,9720, 0,9908, 0,9845, 0,9752, 0,9982 for the second training-test data set and the first CNN architecture, and 0,9733, 0,9920, 0,9857, 0,9772, 0,9987 for the second training-test data set and the second CNN architecture. In this context, it can be seen that the achievements of the first and second CNN architectures are generally close to each other. However, when a comparison is made in terms of CPU run-time, it is understood that the second CNN architecture is two times slower than the first CNN architecture in terms of CPU run-time. The main reason for this is that the number of layers in the second CNN architecture is approximately twice that as high as in the first CNN architecture. A similar situation arose in the experiments performed by combining all the data and using 2-fold cross. For these experiments, the highest mean sensitivity, specificity, accuracy, F-1 score, and AUC values obtained without using the pipeline algorithms are respectively; 0,9253, 0,9892, 0,9719, 0,9468, 0,9939 for the first CNN architecture and 0,9240, 0,9936, 0,9745, 0,9511, 0,9975 for the second CNN architecture. Within the scope of the study, DT-CWT was used to reduce the image dimensions. In this way, DT-CWT tolerated the increase in result-producing time due to the use of the pipeline algorithm. In this context, when the results obtained using the original images and the ones obtained using DT-CWT are compared, it can be seen that there is no serious decrease in the results, in general. Using DT-CWT, the image sizes were reduced successfully and a reduction in the result-producing times was achieved, in the study. The pipeline algorithms proposed within the scope of the study are based on combining the results obtained without using LBP and with using LBP, as detailed previously. After this stage, the study results obtained by using the pipeline algorithms were analyzed. With the introduction of the pipeline algorithms, improvements were achieved in all the parameters obtained by using both training-test data sets and the CNN architectures. In this context, an improvement was achieved in general, according to the highest results obtained without LBP and with using LBP, in terms of percentage ranging between 0,67% and 3,73% for the sensitivity parameter, between 0,06% and 2,25% for the specificity parameter, between 0% to 2,61% for the accuracy parameter, between 0,03% and 2,04% for the F-1 score parameter, and between 0% and 1,20% for the AUC parameter. It was also observed that similar improvements were achieved for the experiments performed by combining all data and using 2-fold cross. In this context, according to the highest results obtained without LBP and with using LBP, an improvement was achieved generally in terms of percentage ranging between 2,13% and 5,07% for the sensitivity parameter, between 0,59% and 1,08% for the specificity parameter, between 0,58% and 1,87% for the accuracy parameter, between 1,18% and 3,55% for the F-1 score parameter, and between 0,13% and 0,59% for the AUC parameter. When comparing the success of pipeline algorithms in improving the results in general, it can be seen that the algorithms of pipeline-1 and pipeline-3 obtain the highest sensitivity values; pipeline-4 obtains the highest specificity values; pipeline-1 and pipeline-3 obtain the highest accuracy values; pipeline-1 and pipeline-3 obtain the highest F-1 scores values; and pipeline-1, pipeline-2 and pipeline-3 algorithms successfully obtained the highest AUC values. When the input data with the best results obtained by using the pipeline algorithms are examined, it can be seen that using the real part of the LL sub-image band for the first training-test data set and using the original images for the second training-test data set provided the best results. Experiments performed using the 2-fold cross by combining all the data also confirm this situation. For this reason, only the results of the experiments mentioned were included in the study, in consideration of the length of the study. The highest mean sensitivity, specificity, accuracy, F-1 score, and AUC values obtained using the study pipeline algorithms are as follows, respectively; 0,9947, 0,9800, 0,9843, 0,9881, 0,9990 for the first training-test data set and the first CNN architecture; 0,9867, 0,9800, 0,9809, 0,9853, 0,9977 for the first training-test data set and the second CNN architecture; 0,9853, 0,9926, 0,9857, 0,9774, 0,9988 for the second training-test data set and the first CNN architecture; and 0,9920, 0,9939, 0,9891, 0,9828, 0,9991 for the second training-test and the second CNN architecture. The highest mean sensitivity, specificity, accuracy, F-1 score and AUC values obtained in the experiments performed by combining all data and using the 2-fold cross were respectively; 0,9760, 1,0000, 0,9906, 0,9823, 0,9997 for the first CNN architecture; and 0,9707, 1,0000, 0,9867, 0,9752, 0,9994 for the second CNN architecture. Within the scope of the study, the best results obtained before and after using the pipeline algorithm and the comparison of these results with the recent literature studies are given in Table 24.
Table 24

Comparison of the results obtained, within the scope of the study, with previous studies

StudySENSPEACCF-1AUC
Tuncer et al. [26]0,8149-1,00000,9380-1,00000,9049-0,9955XX
Panwar et al. [27]0,97620,78570,881XX
Ozturk et al. [28]0,95130,9530,98080,9651X
Mohammed et al. [29]0,706-0,9740,557-1,0000,620-0,9870,555–0,9870,800-0,988
Khan et al. [30]0,9930,9860,9900,985X
Apostolopoulos and Mpesiana [31]0,98660,96460,9678XX
Waheed et al. [32]0,69-0,900,95-0,970,85-0,95XX
Mahmud et al. [33]0,9780,9470,9740,9710,969
Vaid et al. [34]0,98630,91660,96330,9729X
Benbrahim et al. [35]0,9803-0,9811X0,9803-0,99010,9803-0,9901X
Elaziz et al. [36]0,9875-0,9891X0,9609-0,9809XX
Martínez et al. [37]0,97X0,970,97X
Loey et al. [38]1,00001,00001,0000XX
Toraman et al. [39]0,28-0,97420,8095–0,980,4914-0,97240,55-0,9724X
Duran-Lopez et al. [40]0,92530,96330,94430,93140,988
Minaee et al. [41]0,980,751-0,929XXX
Our Study (Before Pipeline-First data set)0,98530,89500,95390,96550,9939
Our Study (Before Pipeline-First data set)0,97470,97250,97390,97990,9976
Our Study (Before Pipeline-First data set)0,98000,97000,97650,98190,9975
Our Study (Before Pipeline-First data set)0,97730,96750,97390,98000,9983
Our Study (Before Pipeline-Second data set)0,97330,99140,98570,97720,9981
Our Study (Before Pipeline-Second data set)0,96400,99200,98320,97300,9985
Our Study (Before Pipeline-Second data set)0,96800,99020,98320,97320,9982
Our Study (Before Pipeline-Combined data set)0,92530,98920,97190,94680,9937
Our Study (Before Pipeline-Combined data set)0,91870,99360,97340,94860,9975
Our Study (Before Pipeline-Combined data set)0,92400,99310,97450,95110,9949
Our Study (After Pipeline-First data set)0,99470,95750,98170,98620,9990
Our Study (After Pipeline-First data set)0,98130,98000,98090,98530,9977
Our Study (After Pipeline-First data set)0,99330,96750,98430,98810,9988
Our Study (After Pipeline-Second data set)0,99200,98770,98910,98280,9991
Our Study (After Pipeline-Second data set)0,96400,99390,98450,97500,9980
Our Study (After Pipeline-Combined data set)0,97600,98920,98560,97340,9984
Our Study (After Pipeline-Combined data set)0,96531,00000,99060,98230,9997
Comparison of the results obtained, within the scope of the study, with previous studies

Discussion

As a result of our study on the automatic classification of chest X-ray images and using one of the deep learning methods, the CNN, some important and comprehensive test results were obtained for early diagnosis of Covid-19 disease. When the results obtained within the scope of the study are compared with the literature studies detailed in Tables 1 and 24, the results of the study were found to be better than the 14 out of the 16 studies in which this value was calculated for the sensitivity parameter, than all the 13 studies in which this value was calculated for the specificity parameter, than the 13 out of the 15 studies in which this value was calculated for the accuracy parameter, than the eight out of the nine studies in which this value was calculated for the F-1 score parameter, and than all the 3 studies in which this value was calculated for the AUC parameter. Moreover, if it is necessary to make a comparison in terms of run-times, it was found that it produced a result at least three times faster in terms of run-time than the result was obtained in the study conducted by Mohammed et al. [29]. This study is the only study in which this parameter was calculated. Also, it is at least ten times faster than the study conducted by Toraman et al. [39]. These two studies were studies in which the run-times were shared. No information was given about run-times in the other previous studies. Overall, the results obtained within the scope of the study lagged behind the results obtained in studies conducted by Tuncer et al. [26], Benbrahim et al. [35], and Loey et al. [38]. However, in order to make a more detailed comparison, the number of images used in these studies should be compared with the number of images used in our study. The number of images used in our study is higher than the number of images used in these studies. In particular, the number of images used in our study is almost three times the number of images used by Loey et al. [38]. Another important issue is the procedure for training and testing. There was no cross validation in the studies by Benbrahim et al. [35] and Loet et al. [38]. In our study, cross-validation in the training-test processes is one of the important measures taken against the overfitting problem that occurs during the training of the network. However, it is known that cross validation improves the reliability of the study results while balancing the study results. In this context, these issues should be taken into consideration when making a comparison. In the context of the study, if an evaluation should be based on the differentiation made between giving the images to the CNN as input directly and after the LBP was applied, it can be seen that the images obtained by applying the LBP produced worse results than the original images. However, the pipeline classification algorithm presented in the context of this study enabled the results obtained to be improved by combining the original and LBP-applied images. In this context, a significant part of the best results obtained in the study was provided using the pipeline classification algorithm. In this sense, it can be seen that the results of the study support some other literature studies [61-66] where the CNN and LBP methods are used together and use of the LBP was shown to increase the success of the relevant study. The success achieved through the pipeline approaches in the study is due to the fact that some classification results that could not be revealed without using the LBP alone and with using the LBP alone were revealed by using the two methods together. Feeding the results from the two sources in the pipeline approaches results in an increase in running time. However, the results obtained within the scope of the study show that this time cost can be eliminated by using DT-CWT. In this way, it has been observed that working success can be increased significantly without time cost. It is considered that this model is within the scope of the study and can be used in many other deep learning studies. It was evaluated that another important factor in achieving the successful results in this study was the framing process, which included the chest region and clarified the area of interest before the training and test procedures started. Hence, thanks to this pre-process carried out in this context, the parts lacking medical diagnostic information were removed from the images and only the relevant areas on the images were used in the procedures. As the size of the inputs given to the CNN increases, the time taken for the training and testing increases. The DT-CWT transformation used in the study reduces the size of the image by half. Although the image sizes are reduced by half, there is no serious adverse effect on the study results. By contrast, some of the best results achieved in the study were obtained using the DT-CWT. In this context, although the pipeline classification algorithms proposed in the study increase the time to produce the results for the image, the times in question are less than half the time required for the images to be used directly without applying LBP and DT-CWT. Also, all the training and test procedures provided in the study reflect the amount per image. However, approximately 98% of these periods are spent on the training procedures. In this context, in the case where the results obtained by the transfer learning approach are used with the pipeline classification algorithm proposed in the study, the periods mentioned will decrease accordingly. The pipeline algorithms revealed within the scope of the study were tested for data sets with different weights in terms of the number of Covid-19 and non-Covid-19 images, for different training-test ratios and different CNN architectures. The pipeline algorithms were successful for all these situations that may have affected the results. This shows that the proposed pipeline algorithms are not partial but are general solutions. From this point of view, it is obvious that if the pipeline algorithms mentioned above are added to the algorithms used in other literature studies, this would increase the success of these studies. The results of the study show that analyzing chest X-ray images in the diagnosis of Covid-19 disease using deep learning methods will speed up the diagnosis and significantly reduce the burden on healthcare personnel. To further improve the results of the study, increasing the number of images in the training set, i.e., the creation of databases in which the clinical data of patients with Covid-19 that are accessible to the public, is of prime importance. After this stage, it is aimed to realize applications using CT images of the lungs an important diagnostic tool, such as chest X-ray images, in Covid-19 disease diagnosis. In addition, it is planned to analyze the effects of using the results obtained, through direct transfer learning in pipeline classification algorithms, on the study results. This is evaluated as another important application to classify the complex-valued sub-bands of images obtained by applying DT-CWT, with the help of using the complex-valued CNN directly.
  38 in total

1.  Clinical characteristics of 140 patients infected with SARS-CoV-2 in Wuhan, China.

Authors:  Jin-Jin Zhang; Xiang Dong; Yi-Yuan Cao; Ya-Dong Yuan; Yi-Bin Yang; You-Qin Yan; Cezmi A Akdis; Ya-Dong Gao
Journal:  Allergy       Date:  2020-02-27       Impact factor: 13.146

2.  Two public chest X-ray datasets for computer-aided screening of pulmonary diseases.

Authors:  Stefan Jaeger; Sema Candemir; Sameer Antani; Yì-Xiáng J Wáng; Pu-Xuan Lu; George Thoma
Journal:  Quant Imaging Med Surg       Date:  2014-12

3.  A novel comparative study for detection of Covid-19 on CT lung images using texture analysis, machine learning, and deep learning methods.

Authors:  Huseyin Yasar; Murat Ceylan
Journal:  Multimed Tools Appl       Date:  2020-10-06       Impact factor: 2.757

4.  CovidGAN: Data Augmentation Using Auxiliary Classifier GAN for Improved Covid-19 Detection.

Authors:  Abdul Waheed; Muskan Goyal; Deepak Gupta; Ashish Khanna; Fadi Al-Turjman; Placido Rogerio Pinheiro
Journal:  IEEE Access       Date:  2020-05-14       Impact factor: 3.367

5.  Imaging Profile of the COVID-19 Infection: Radiologic Findings and Literature Review.

Authors:  Ming-Yen Ng; Elaine Y P Lee; Jin Yang; Fangfang Yang; Xia Li; Hongxia Wang; Macy Mei-Sze Lui; Christine Shing-Yen Lo; Barry Leung; Pek-Lan Khong; Christopher Kim-Ming Hui; Kwok-Yung Yuen; Michael D Kuo
Journal:  Radiol Cardiothorac Imaging       Date:  2020-02-13

6.  A Locally Transmitted Case of SARS-CoV-2 Infection in Taiwan.

Authors:  Ying-Chu Liu; Ching-Hui Liao; Chin-Fu Chang; Chu-Chung Chou; Yan-Ren Lin
Journal:  N Engl J Med       Date:  2020-02-12       Impact factor: 91.245

7.  Clinical Characteristics of Imported Cases of Coronavirus Disease 2019 (COVID-19) in Jiangsu Province: A Multicenter Descriptive Study.

Authors:  Jian Wu; Jun Liu; Xinguo Zhao; Chengyuan Liu; Wei Wang; Dawei Wang; Wei Xu; Chunyu Zhang; Jiong Yu; Bin Jiang; Hongcui Cao; Lanjuan Li
Journal:  Clin Infect Dis       Date:  2020-07-28       Impact factor: 9.079

8.  First Case of 2019 Novel Coronavirus in the United States.

Authors:  Michelle L Holshue; Chas DeBolt; Scott Lindquist; Kathy H Lofy; John Wiesman; Hollianne Bruce; Christopher Spitters; Keith Ericson; Sara Wilkerson; Ahmet Tural; George Diaz; Amanda Cohn; LeAnne Fox; Anita Patel; Susan I Gerber; Lindsay Kim; Suxiang Tong; Xiaoyan Lu; Steve Lindstrom; Mark A Pallansch; William C Weldon; Holly M Biggs; Timothy M Uyeki; Satish K Pillai
Journal:  N Engl J Med       Date:  2020-01-31       Impact factor: 91.245

9.  2019-novel Coronavirus severe adult respiratory distress syndrome in two cases in Italy: An uncommon radiological presentation.

Authors:  Fabrizio Albarello; Elisa Pianura; Federica Di Stefano; Massimo Cristofaro; Ada Petrone; Luisa Marchioni; Claudia Palazzolo; Vincenzo Schininà; Emanuele Nicastri; Nicola Petrosillo; Paolo Campioni; Petersen Eskild; Alimuddin Zumla; Giuseppe Ippolito
Journal:  Int J Infect Dis       Date:  2020-02-26       Impact factor: 3.623

10.  CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest x-ray images.

Authors:  Asif Iqbal Khan; Junaid Latief Shah; Mohammad Mudasir Bhat
Journal:  Comput Methods Programs Biomed       Date:  2020-06-05       Impact factor: 5.428

View more
  4 in total

1.  Machine Learning-Based Research for COVID-19 Detection, Diagnosis, and Prediction: A Survey.

Authors:  Yassine Meraihi; Asma Benmessaoud Gabis; Seyedali Mirjalili; Amar Ramdane-Cherif; Fawaz E Alsaadi
Journal:  SN Comput Sci       Date:  2022-05-12

2.  A novel study for automatic two-class COVID-19 diagnosis (between COVID-19 and Healthy, Pneumonia) on X-ray images using texture analysis and 2-D/3-D convolutional neural networks.

Authors:  Huseyin Yaşar; Murat Ceylan
Journal:  Multimed Syst       Date:  2022-01-29       Impact factor: 2.603

3.  Real-time echocardiography image analysis and quantification of cardiac indices.

Authors:  Ghada Zamzmi; Sivaramakrishnan Rajaraman; Li-Yueh Hsu; Vandana Sachdev; Sameer Antani
Journal:  Med Image Anal       Date:  2022-06-09       Impact factor: 13.828

4.  Automatic coronavirus disease 2019 diagnosis based on chest radiography and deep learning - Success story or dataset bias?

Authors:  Jennifer Dhont; Cecile Wolfs; Frank Verhaegen
Journal:  Med Phys       Date:  2022-01-12       Impact factor: 4.506

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.