Literature DB >> 34345118

Improving effectiveness of different deep learning-based models for detecting COVID-19 from computed tomography (CT) images.

Erdi Acar1, Engin Şahin1, İhsan Yılmaz1.   

Abstract

COVID-19 has caused a pandemic crisis that threatens the world in many areas, especially in public health. For the diagnosis of COVID-19, computed tomography has a prognostic role in the early diagnosis of COVID-19 as it provides both rapid and accurate results. This is crucial to assist clinicians in making decisions for rapid isolation and appropriate patient treatment. Therefore, many researchers have shown that the accuracy of COVID-19 patient detection from chest CT images using various deep learning systems is extremely optimistic. Deep learning networks such as convolutional neural networks (CNNs) require substantial training data. One of the biggest problems for researchers is accessing a significant amount of training data. In this work, we combine methods such as segmentation, data augmentation and generative adversarial network (GAN) to increase the effectiveness of deep learning models. We propose a method that generates synthetic chest CT images using the GAN method from a limited number of CT images. We test the performance of experiments (with and without GAN) on internal and external dataset. When the CNN is trained on real images and synthetic images, a slight increase in accuracy and other results are observed in the internal dataset, but between 3 % and 9 % in the external dataset. It is promising according to the performance results that the proposed method will accelerate the detection of COVID-19 and lead to more robust systems.
© The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2021.

Entities:  

Keywords:  COVID-19; Computed tomography; Data augmentation; Deep learning; Generative adversarial network; Lung segmentation

Year:  2021        PMID: 34345118      PMCID: PMC8321007          DOI: 10.1007/s00521-021-06344-5

Source DB:  PubMed          Journal:  Neural Comput Appl        ISSN: 0941-0643            Impact factor:   5.102


Introduction

COVID-19 is an infectious disease with a high mortality rate caused by the mutant SARS-CoV-2 virus. Following the identification of the first case in Wuhan, China, in December 2019, it quickly spread all over the world and was declared a pandemic by WHO on March 11, 2020. COVID-19 has caused major public health problems in the international community due to its rapid spread all over the world [1]. It is extremely important to be able to diagnose COVID-19 in an infected patient during the pandemic process. Although polymerase chain reaction (PCR) testing is the standard for confirming COVID-19 positive patients, medical imaging such as X-ray and non-contrast computed tomography (CT) plays an important role in the detection of COVID-19. Since COVID-19 can be detected in the early stages with the presence of lung ground-glass opacities in CT images, which are clearer and more accurate than X-ray images, the diagnosis of COVID-19 from CT will help clinicians make quick decisions for rapid isolation and appropriate patient treatment [2]. Recently, computer vision has led to extraordinary developments with the advancement of artificial intelligence technology and especially the development of convolutional neural networks (CNN). It is widely used in the medical field and provides support for medical diagnosis [3-5]. Additionally, due to the heavy workload of large numbers of infected patients and healthcare professionals during the pandemic crisis, the AI-based computer-aided system could speed up the diagnostic process. These systems can provide effective diagnosis of the disease in a shorter time in cases such as radiologist insufficiency and deficiency. It can also reduce the need for human surveillance and identify details invisible to the human eye. In this context, many deep learning models have also begun to be developed for the diagnosis of COVID-19. There is currently a great deal of interest in data-driven approaches to diagnosing COVID-19. Many researchers have used deep convolutional neural networks (CNN) such as VGG, ResNet, Inception, Xception and DenseNet, with well-known and proven performance, to diagnose COVID-19. However, deep learning networks such as convolutional neural networks require large amounts of training data. In addition, because CNNs have a large number of parameters, overfitting can easily occur in small datasets. However, medical imaging data are scarce, expensive and fraught with legal concerns over patient privacy, so datasets are small in size. A quick and easy method, data augmentation, can be used to overcome this problem [6]. The dataset is artificially expanded using techniques such as image rotation, transformation, scaling, brightness or contrast enhancement. However, it cannot be said that completely different images are produced since the same data are displayed differently with this method. In addition, the dataset can be enlarged by generating unique data by producing synthetic data with GAN, which is an advanced technique [7, 8]. The main contributions of this article can be summarized as follows:The rest of the paper is organized as follows. In Sect. 2, brief information about the studies related to our study is given. The materials and methods used in the study are presented in Sect. 3. Section 4 presents the results of different analyzes for different deep learning algorithms in the proposed frameworks on internal and external datasets. The comparisons with the other related studies and the discussion of the results are given in Sect. 5. Section 6 presents the conclusion of this study. It is difficult to identify COVID-19 as findings on CT images differ in both position and shape in different patients due to infections caused by COVID-19. In addition, we present a method for generating synthetic CT images by developing a GAN-based model to overcome the problem of overfitting in CNNs. In addition, we demonstrate the effectiveness of GAN-generated synthetic images in improving CNN’s generalization ability for COVID-19 detection in the context of performance criteria in internal and external datasets. We use a segmentation model based on ConvLSTMU-Net architecture and graph-cut image processing for segmentation of the relevant lung region from CT images. We use data enlargement techniques such as random distortion, rotation, flip and zoom with an in-place / on-the-fly augmentation approach to increase the effectiveness of the model during model training.

Related works

Recently, many studies have been conducted to perform data augmentation on medical images. Zhoe et al. [9] developed the forward and backward GAN (F&BGAN) method to create synthetic images of lung nodules in lung cancer and used the VGG16 network for benign and malignant classification. Chuquicusma et al. [10] produced realistic lung nodules using DCGAN for the training of radiologists. Frid-Adar et al. [11] proposed a magnification scheme for advanced liver lesion classification based on a combination of standard image perturbation and synthetic liver lesion generation using GAN. Guibas et al. [12] proposed a two-stage pipeline for generating synthetic medical images from a pair of GANs, which was tested in practice on retinal fundi images. Shin et al. [13] proposed a method to generate synthetic abnormal MRI images with brain tumors by training a GAN using two publicly available datasets of brain MRI. Nowadays, there is intense interest in diagnosing COVID-19 using deep learning methods. Wu et al. [14] aimed to diagnose COVID-19 cases with the concept of multi-view fusion using the ResNet50 architecture. The datasets they used in their studies included 622 CT images collected from two different hospitals in China. They resized their images in the datasets to . The system they developed achieved accuracy, sensitivity and specificity. Xu et al. [15] collected the data they used in their study from 3 different hospitals in China. It achieved accuracy, sensitivity, accuracy and F1-score with pretrained ResNet18 architecture using 618 CT images. Jin et al. [16] used pretrained CNN models such as DPN-92, Inceptionv3, ResNet50 and Attention ResNet50 with 3D U-Net++ for the diagnosis of COVID-19. The datasets contain 850 COVID-19 positives and 541 negative COVID-19 images and were collected from 5 different hospitals in China. According to the performance criteria, their system accepted the 3D U-Net++ResNet50 model as the best model with accuracy, sensitivity and AUC values. Yousefzadeh et al. [17] introduced a deep learning framework using different CNN architectures DenseNet, ResNet, Xception and EfficientNetB0. The datasets contain a total of 2124 CT images. The system they proposed resulted in accuracy, sensitivity, specificity, F1 score and AUC. Ardakani et al. [18] proposed a system for detecting COVID-19 using CNN architectures such as AlexNet, VGG-16, VGG-19, SqueezeNet, GoogleNet, MobileNetv2, ResNet18, ResNet50, ResNet101 and Xception. The dataset they use contains 1020 CT images in total. Among the 10 networks, ResNet101 and Xception performed better than the others according to their performance values. Chen et al. [19] used the U-Net++ segmentation model together with the ResNet50 model to diagnose COVID-19 using a very large and non-public dataset of 106 patients from the Renmin Hospital of Wuhan University. The results of COVID-19 classification are accuracy, sensitivity and specificity. The U-Net and 3D CNN models are used for lung segmentation and diagnosing of COVID-19, respectively, in Ref. [20]. The results of the model are sensitivity, specificity and AUC. Most of the previous works on COVID-19 are given in Refs. [21-23].

Material and methods

Dataset

We use two different datasets in our study: The dataset, which was collected from different open sources [24], contains 1607 COVID-19 and 1667 normal CT images. We randomly allocate 80% of the dataset for training and 20% for testing. We will refer to this dataset as the internal dataset throughout the study. The external dataset in Ref. [25] contains 4001 COVID-19 and 9575 normal CT images. We use this dataset to evaluate the performance of the trained models.

Preprocessing

Since the internal dataset is collected from different sources, it contains CT images with a resolution of min - max . For these reasons, the images in the internal dataset were resized to resolution. In the external dataset, all CT images are high-resolution images. We also reduce the size of the CT images in this dataset to to be standard and perform central zooming.

Lung segmentation

Since distinctive information is in the lungs on CT images, we apply lung segmentation to obtain only the lung area. For segmentation, we applied a similar approach to BidirectionalConvLSTMU-Net (BDCU-Net) [26] architecture. A total of 1667 normal CT data and masks in the internal dataset were used for training the lung segmentation architecture. We reserve of the dataset for training and of the dataset for validation. We initially started with learning rate and training epochs of 30. We dynamically reduced the learning rate when there was no improvement in validation error for 4 consecutive training periods. When there is no improvement in validation error for 8 consecutive training epochs, it is taken as the criterion for early stopping of training. The Adam stochastic optimization algorithm, which was developed as a solution to the disappearing gradient problem, was used for parameter optimization. At the end of the 26th epoch, an early stop occurs, and accuracy: 0.9675 is obtained. The architecture of obtaining the mask images from the CT images by using lung segmentation model is given in Fig. 1. The sample images and obtained mask images by using the lung segmentation model are given in Fig. 2. The example images for the relevant region obtained as a result of Graphcut image processing [27] are given in Fig. 3.
Fig. 1

The architecture of obtaining the mask images from the CT images by using BDCU-Net model

Fig. 2

Sample CT images and mask images obtained as a result of BDCU-Net

Fig. 3

Sample images for the region of interest obtained after applying Graphcut image processing

The architecture of obtaining the mask images from the CT images by using BDCU-Net model Sample CT images and mask images obtained as a result of BDCU-Net Sample images for the region of interest obtained after applying Graphcut image processing

Generating synthetic images

Generative adversarial network (GAN)

The GAN framework is a widely used modern method for generating synthetic data from many domains [28, 29]. Examples of data generation with GANs include text-to-image synthesis, superimage resolution, style transfer and symbolic music production [30]. The working principle of GANs is simply; GAN is a min-max game between two hostile networks, generator G(z) and discriminator D(x). The generator tries to convert the random noise into observations that appear to be sampled from the original dataset, and the discriminator tries to guess whether an observation comes from the original dataset or is one of the generator’s fakes. At the start of the process, generator G(z) takes a z-point (random noise) from a latent space and outputs noisy images and tries to fool the discriminator D(x). If the discriminator is D(x), the generator tries to distinguish whether the images from are real or fake. The key of GANs lies in how we change the training of the two networks so that as the generator becomes more adept at deceiving the discriminator, the discriminator must adapt to maintain its ability to accurately identify which observations are fake. Thus, the generator tries to find new ways to fool the discriminator so that the loop continues between the two networks. The purpose of G(z) is to minimize the cost function V(D, G) and to maximize the D(x) by training both G(z) and D(x) at the same time:The architecture of GAN applied in the study is given in Fig. 4
Fig. 4

The framework of GAN

The framework of GAN

GAN Train Procedure

Initially, the parameters of the discriminator network are set to non-trainable. The output of the generator network feeds the discriminator and the generator is updated through the discriminator, so the generator is stacked on the discriminator. The hyperparameter of LeakyReLu layers used in discriminator and generator networks is , and momentum value of batch normalization layer used in generator is 0.8. The detailed designs of the discriminator and generator network used in the study are given in Fig. 5.
Fig. 5

The detailed designs of the discriminator and generator network used in the study

The detailed designs of the discriminator and generator network used in the study of the internal dataset reserved for training iz used in GAN training, i.e., 1286 COVID-19 and 1333 normal CT images. Test data are not included in the training. The pixel values of the CT images are normalized from [0, 255] to as a preprocessing. In the training of both networks, Adam optimization algorithm with learning rate and momentum value is used as hyperparameters. The least-squares function is used as the loss function. In addition, the number of training periods is determined as 40000 and the batch size is 64. The latent space dimension where the generator is fed, is taken as 100. A total of 2314 normal and 2267 COVID-19 synthetic images are generated by GAN. The samples of the COVID-19 and normal synthetic images produced after the GAN training are given in Fig. 6.
Fig. 6

a The samples of the COVID-19 synthetic images produced after the training with GAN b The samples of the normal synthetic images produced after the training with GAN

a The samples of the COVID-19 synthetic images produced after the training with GAN b The samples of the normal synthetic images produced after the training with GAN

CNN architectures

Deep CNN architectures pretrained on 9 well-known ImageNet [31] datasets were used in this study: VGG16, VGG19, Xception, ResNet50, ResNet50v2, Inceptionv3, InceptionResNetv2, DenseNet121 and DenseNet169. VGG16 was developed to win the ILSVRC2014 (Large Scale Visual Recognition Challenge 2014) in 2014 [32]. VGG16 is a convolutional neural network architecture named after the Visual Geometry Group of Oxford developed. VGG-16 is a convolutional neural network with a depth of 16 layers. The model loads a pretrained set of weights on ImageNet and reaches accuracy on ImageNet, a dataset of more than 14 million images belonging to 1000 classes. VGG-19 is the version of VGG with a depth of 19 layers. VGG16 has 138, 357, 544 parameters while VGG19 has 143, 667, 240 parameters. He et al. [33] developed the architecture called residual network (ResNet) to solve the vanishing/exploding gradient problem. ResNet is created by adding the blocks that feed the residual values to the next layers into the model. This is also called residual learning. ResNet-50 has 16 residual blocks. The main difference between ResNet50 and ResNet50v2 is now the arrangement of layers in blocks. ResNet50 and ResNet50v2 achieved and accuracy on the ImageNet dataset, respectively. Inception [34] and InceptionResNet [35] are developed by Google. The concept of Inception represents Inception modules within the network. This is a simple and powerful architectural unit that also allows the model to learn parallel filters of different sizes and multiple scales. The main difference between Inception and InceptionResNet is that InceptionResNet uses residual layers. Inceptionv3 using the Inception modules has 23, 851, 784 while InceptionResNet has 55, 873, 736. Inceptionv3 and InceptionResNetv2 achieved and accuracy on the ImageNet dataset, respectively. Xception [36] is a deep convolutional neural network architecture that includes deeply separable convolutions and developed by Google researchers. Google presented an interpretation of the Initiation modules in convolutional neural networks as an intermediate step between regular convolution and deeply separable convolution processing. Xception has 22, 910, 480 parameters and achieved accuracy on the ImageNet dataset. DenseNet [37] is a type of convolutional neural network that uses dense connections between layers through dense blocks that connect all layers directly. Each layer receives additional inputs from all previous layers and transmits its own feature maps to all subsequent layers. The difference between DenseNet121 and DenseNet169 is the number of layers. DenseNet121 has 8, 062, 504 parameters while DenseNet169 has 14, 307, 880 parameters. DenseNet121 and DenseNet169 achieved and accuracy on the ImageNet dataset, respectively.

Training of models and implementation details

In this study, two types of experimental studies (with GAN and without GAN) are conducted using the real images from the internal dataset and incorporating synthetic images generated by the GAN into the internal dataset. Details of the dataset used in the two studies are given in Table 1. The 1667 normal class images in the internal dataset used in the training of models are the same as the images used in the BConvLSTM U-Net training.
Table 1

The numbers of the real images without GAN and the real+synthetic images produced with GAN in training and test sets

FrameworkTypesTraining sizeTest size
Real imagesCOVID-191286321
Without GANNormal1333334
Real+synthetic imagesCOVID-193600321
Produced with GANNormal3600334
The numbers of the real images without GAN and the real+synthetic images produced with GAN in training and test sets All networks in both frameworks, the pixel values of the CT images, were normalized from [0, 255] to [0, 1] as a preprocessing. In both experiments, Adam optimization algorithm with learning rate and momentum value was used as hyperparameter. Binary cross-entropy function is used as loss function. In addition, the number of training epochs is determined as 40 and the batch size was 16. The learning rate is reduced when there is no improvement in validation error for 5 consecutive training periods. It was taken as the criterion for early discontinuation of training when there is no improvement in validation error for 10 consecutive training periods. A fivefold cross-validation is performed and tested on a randomly selected of the internal dataset and the external dataset. In both experiments, pretrained deep CNN models VGG16, VGG19, Xception, ResNet50, ResNet50V2, DenseNet121, DenseNet169, InceptionV3 and InceptionResNetV2 were used and fine-tuned. Dense(512, relu), BatchNormalization (0.9), Dense(256, relu), Dense (1, Sigmoid) layers are used, respectively, in the fully connected layer for classification. The working principle of this study is given in Fig. 7
Fig. 7

The flow chart of both experiments is given in Fig. 5

The flow chart of both experiments is given in Fig. 5

Data augmentation

The performance of deep learning neural networks is generally directly proportional to the amount of data available. Data augmentation is referred to as a technique of artificially deriving data from existing training data through much more image processing such as rotating, panning, zooming, flipping [6]. The aim is to increase the generalization performance of the model while expanding the training dataset with new examples. Thus, the trained model constantly sees new, different versions of the input data, and the model can learn more robust features. There are three types of data augmentation methods: first, expanding the existing dataset. The problem with this approach is that it does not fully increase the generalization ability of the model. The second is in-place/on-the-fly data augmentation. The network is trained to see new variations of the data in each epoch by using this type of data augmentation. This method increases the generalization ability of the model. Finally, a hybrid approach combines the first and the second approaches. In our study, we use only in-place/on-the-fly augmentation approach in models trained with real data; in the second experiment, we increase the size of the dataset with synthetic data produced with GAN, and we adopt a hybrid approach with in-place/on-the-fly augmentation during the training of models. The applied data augmentation processes and sample images are given in Table 2 and Fig. 8, respectively. Random distortion allows you to distort the image while maintaining elastically the image aspect ratio. The magnitude parameter determines how much it degrades. Flip left-right acts as a mirror on images. Rotation was used to rotate left and right a certain degree. Zoom was used to enlarge and reduce the image centrally.
Table 2

Types of data augmentation

TypesParameters
Random distortionprobability=0.5
Grid width=4
Grid height=4
Magnitude=10
Flip (left, right)Probability=0.5
RotateProbability=0.5
Max left rotation=10
Max right rotation=10
ZoomProbability=0.5
Min factor=0.9
Max factor=1.20
Fig. 8

a Original image b result of the flip left-right c result of the right rotation d result of the left rotation e result of a 0.9 rate zoom f result of a 1.20 rate zoom. g and h results of random distortion

Types of data augmentation a Original image b result of the flip left-right c result of the right rotation d result of the left rotation e result of a 0.9 rate zoom f result of a 1.20 rate zoom. g and h results of random distortion

Results

All training and testing processes are performed using AMD Ryzen 3970X CPU with 128GB RAM and Nvidia RTX 3080 GPU with 10GB memory. The Keras [38] deep learning library is used for processes. Estimation performances of the methods in this study are measured with metrics such as accuracy, precision, recall and F1-score. The confusion matrix used to explain the performance of the classification model consists of true positive (), true negative (TN), false positive (FP) and false negatives (FN). Dividing the number of correctly classified cases into the total number of test images shows the accuracy and is calculated as follows.where is the number of instances that correctly predicted, is the number of instances that incorrectly predicted, is the number of negative instances that correctly predicted, is the number of negative instances that incorrectly predicted. The recall is used to measure correctly classified COVID-19 cases. Recall is calculated as follows.The percentage of correctly classified labels in truly positive patients is defined as the precision and is calculated as follows.Specificity is the proportion of people who test negatively among those who actually do not have the disease and is defined as:The F1-score is defined as the weighted average of precision and recall combining both precision and recall and is calculated as follows.The same real images are used in the tests of the both experiments on both internal and external datasets. Information on test sets size is given in Table 3. The comparison of performance results in internal and external datasets for both experiments is given in Tables 4 and 5, respectively. Confusion values and performance results for each model are given in Section Appendix.
Table 3

Information on the size of the test sets

DatasetTypesTesting set
InternalCOVID-19321
Normal334
ExternalCOVID-199545
Normal4001
Table 4

Comparison performance results of models without GAN and with GAN on internal dataset

ModelMethodAccuracyPrecisionRecallSpecificityF1-score
VGG16without GAN0.98140.98980.97200.99040.9808
with GAN0.98720.99490.97880.99520.9868
VGG19without GAN0.98410.99240.97510.99280.9837
with GAN0.98750.98940.98500.98980.9872
Xceptionwithout GAN0.99020.99000.99000.99040.9900
with GAN0.99270.99130.99380.99160.9925
ResNet50without GAN0.97980.98190.97690.98260.9794
with GAN0.99110.99690.98500.99700.9909
ResNet50v2without GAN0.98600.99310.97820.99340.9856
with GAN0.99420.99750.99070.99760.9941
InceptionV3without GAN0.98810.99490.98070.99520.9878
with GAN0.99510.99490.99130.99520.9950
InceptionResNetV2without GAN0.98930.99310.98500.99340.9890
with GAN0.98930.99620.99130.99640.9938
DenseNet121without GAN0.98930.99190.98630.99220.9891
with GAN0.99210.99190.99190.99220.9919
DenseNet169withoutGAN0.99110.99440.98750.99460.9909
with GAN0.99270.99560.98940.99580.9925
Table 5

Comparison performance results of models without GAN and with GAN on external dataset

ModelMethodAccuracyPrecisionRecallSpecificityF1-score
VGG16without GAN0.86700.74470.83640.87990.7879
with GAN0.90230.80240.88770.90830.8429
VGG19without GAN0.88360.76460.87580.88680.8194
with GAN0.90860.80860.90480.91020.8540
Xceptionwithout GAN0.91450.80390.93990.90390.8666
with GAN0.94860.87790.95930.94400.9168
ResNet50without GAN0.89780.79000.89060.90070.8373
with GAN0.93030.86440.90610.94040.8847
ResNet50v2without GAN0.89140.77550.89020.89190.8289
with GAN0.91670.82720.90770.92050.8656
InceptionV3without GAN0.90610.80750.89550.91050.8492
with GAN0.94980.89360.94230.95300.9173
InceptionResNetV2without GAN0.89830.79280.88750.90280.8374
with GAN0.93730.86820.92870.94090.8974
DenseNet121without GAN0.90000.79530.89070.90390.8403
with GAN0.93460.86730.92030.94070.8930
DenseNet169withoutGAN0.90320.79820.89970.90460.8459
with GAN0.93190.86240.91560.93880.8882
Information on the size of the test sets Comparison performance results of models without GAN and with GAN on internal dataset Comparison performance results of models without GAN and with GAN on external dataset

Discussion

In this paper, two experiments with and without GAN were tested on two different datasets (655 internal and 14545 external) with different deep learning methods. Table 4 analyzes the performance results of models with and without GAN on internal dataset. We clearly see that in all deep learning models with GAN, a slightly better performance ratio is achieved in all of the accuracy, precision, recall, specificity and F1-score values compared to the models without GAN. The best detection accuracy rate obtained in the InceptionV3 model with GAN on the internal dataset is and the F1-score rate is . Table 5 analyzes the performance results of models with and without GAN on different external dataset that the system has never seen before. We clearly see that in all deep learning models with GAN, a better performance rate of to is achieved in all of the accuracy, precision, recall, specificity and F1-score values compared to the models without GAN. The best detection accuracy rate obtained in the InceptionV3 model with GAN on the external dataset is and the F1-score rate is . The results of this study reveal that synthetic augments produced with GAN make an important contribution to the improvement of generalization performance, especially in external dataset.

Conclusion

In this research, we developed a GAN-based model to overcome the overfitting problem in CNNs and improve their generalization performance and proposed a model in which the dataset can be artificially amplified with synthetic CT images. An internal dataset containing 3274 CT images was used for training two different experiments (with and without GAN). The improvement in the generalization ability of CNN models was investigated using synthetic data augmentation technique with GAN on this limited dataset. Both experiments were tested on an external dataset. This study uses two classes (COVID-19 and normal) for classification. Synthetic data augmentation with GAN expands the dataset, providing more variability. When the CNN is trained on real images and synthetic images, a slight increase in accuracy and the other results is observed in the internal dataset, but between and in the external dataset. Performance results show that synthetic images produced with GAN make a significant contribution to the detection of COVID-19, especially in external dataset that the system has not seen before. In conclusion, we proposed a method to increase the accuracy of COVID-19 detection with minimal dataset by producing synthetic images of CT images. The proposed method will improve the generalization performance of CNNs and lead to more robust systems.
Table 6

Confusion values of VGG16 on internal dataset

KFoldTPFNTNFP
Without GANFold1311103313
Fold231473322
Fold331383304
Fold4310113313
Fold531293313
With GANFold131563313
Fold231473322
Fold331373321
Fold431273321
Fold531273311
Table 7

Performance results of VGG16 on internal dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.97870.98730.96880.98810.9780
Fold20.98630.99370.97820.99400.9859
Fold30.98170.98740.97510.98800.9812
Fold40.97890.99040.96570.99100.9779
Fold50.98170.99050.97200.99100.9811
Overall0.98440.98980.97200.99040.9808
With GANFold10.98630.99060.98130.99100.9859
Fold20.98630.99370.97820.99400.9859
Fold30.98780.99680.97820.99700.9874
Fold40.98780.99680.97820.99700.9874
Fold50.98780.99680.97820.99700.9874
Overall0.98720.99490.97880.99520.9868
Table 8

Confusion values of VGG19 on internal dataset

KFoldTPFNTNFP
Without GANFold131293331
Fold231383322
Fold331473322
Fold431653313
Fold531563322
With GANFold131563304
Fold231563304
Fold331743313
Fold431743313
Fold531743313
Table 9

Performance results of VGG19 on internal dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.98470.99680.97200.99700.9842
Fold20.98470.99370.97510.99400.9843
Fold30.98630.99370.97820.99400.9859
Fold40.98780.99060.98440.99100.9875
Fold50.98780.99370.98130.99400.9875
Overall0.98630.99370.97820.99400.9859
With GANFold10.98470.98750.98130.98800.9844
Fold20.98470.98750.98130.98800.9844
Fold30.98930.99060.98750.99100.9891
Fold40.98930.99060.98750.99100.9891
Fold50.98930.99060.98750.99100.9891
Overall0.98750.98940.98500.98980.9872
Table 10

Confusion values of Xception on internal dataset

KFoldTPFNTNFP
Without GANFold131743322
Fold231833313
Fold331833304
Fold431833304
Fold531833313
With GANFold131923322
Fold231923313
Fold331923313
Fold431923313
Fold531923313
Table 11

Performance results of Xception without GAN on internal dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.99080.99370.98750.99400.9906
Fold20.99080.99070.99070.99100.9907
Fold30.98930.98760.99070.98800.9891
Fold40.98930.98760.99070.98800.9891
Fold50.99080.99070.99070.99100.9907
Overall0.99020.99000.99000.99040.9900
With GANFold10.99390.99380.99380.99400.9938
Fold20.99240.99070.99380.99100.9922
Fold30.99240.99070.99380.99100.9922
Fold40.99240.99070.99380.99100.9922
Fold50.99240.99070.99380.99100.9922
Overall0.99270.99130.99380.99160.9925
Table 12

Confusion values of ResNet50 on internal dataset

KFoldTPFNTNFP
Without GANFold1311103259
Fold231473268
Fold331563295
Fold431383304
Fold531563313
With GANFold131743322
Fold231743331
Fold331743340
Fold431563331
Fold531563331
Table 13

Performance results of ResNet50 on internal dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.97100.97190.96880.97310.9704
Fold20.97710.97520.97820.97600.9767
Fold30.98320.98440.98130.98500.9828
Fold40.98170.98740.97510.98800.9812
Fold50.98630.99060.98130.99100.9859
Overall0.97980.98190.97690.98260.9794
With GANFold10.99080.99370.98750.99400.9906
Fold20.99240.99690.98750.99700.9922
Fold30.99391.00000.98751.00000.9937
Fold40.98930.99680.98130.99700.9890
Fold50.98930.99680.98130.99700.9890
Overall0.99110.99690.98500.99700.9909
Table 14

Confusion values of ResNet50v2 on internal dataset

KFoldTPFNTNFP
Without GANFold131293331
Fold231293331
Fold331563313
Fold431473313
Fold531743313
With GANFold131743322
Fold231743331
Fold331923331
Fold431833340
Fold531923340
Table 15

Performance results of ResNet50v2 on internal dataset

KFoldAccuracyPrecisionRecall (Sensitivity)SpecificityF1-score
Without GANFold10.98470.99680.97200.99700.9842
Fold20.98470.99680.97200.99700.9842
Fold30.98630.99060.98130.99100.9859
Fold40.98470.99050.97820.99100.9843
Fold50.98930.99060.98750.99100.9891
Overall0.98600.99310.97820.99340.9856
With GANFold10.99080.99370.98750.99400.9906
Fold20.99240.99690.98750.99700.9922
Fold30.99540.99690.99380.99700.9953
Fold40.99541.00000.99071.00000.9953
Fold50.99691.00000.99381.00000.9969
Overall0.99420.99750.99070.99760.9941
Table 16

Confusion values of InceptionV3 on internal dataset

KFoldTPFNTNFP
Without GANFold131563340
Fold231563322
Fold331563322
Fold431563322
Fold531473322
With GANFold132103331
Fold231833331
Fold331833340
Fold431653340
Fold531833340
Table 17

Performance results of InceptionV3 on internal dataset

KFoldAccuracyPrecisionRecall (Sensitivity)SpecificityF1-score
Without GANFold10.99081.00000.98131.00000.9906
Fold20.98780.99370.98130.99400.9875
Fold30.98780.99370.98130.99400.9875
Fold40.98780.99370.98130.99400.9875
Fold50.98630.99370.97820.99400.9859
Overall0.98810.99490.98070.99520.9878
With GANFold10.99850.99691.00000.99700.9984
Fold20.99390.99690.99070.99700.9938
Fold30.99541.00000.99071.00000.9953
Fold40.99241.00000.98441.00000.9922
Fold50.99541.00000.99071.00000.9953
Overall0.99510.99490.99130.99520.9950
Table 18

Confusion values of InceptionResNetV2 on internal dataset

KFoldTPFNTNFP
Without GANFold131383340
Fold231833313
Fold331743322
Fold431653313
Fold531743313
With GANFold131923322
Fold231833331
Fold331833331
Fold431833331
Fold531833331
Table 19

Performance results of InceptionResNetV2 on internal dataset

KFoldAccuracyPrecisionRecall (Sensitivity)SpecificityF1-score
Without GANFold10.98781.00000.97511.00000.9874
Fold20.99080.99070.99070.99100.9907
Fold30.99080.99370.98750.99400.9906
Fold40.98780.99060.98440.99100.9875
Fold50.98930.99060.98750.99100.9891
Overall0.98930.99310.98500.99340.9890
With GANFold10.99390.99380.99380.99400.9938
Fold20.99390.99690.99070.99700.9938
Fold30.99390.99690.99070.99700.9938
Fold40.99390.99690.99070.99700.9938
Fold50.99390.99690.99070.99700.9938
Overall0.98930.99620.99130.99640.9938
Table 20

Confusion values of DenseNet121 on internal dataset

KFoldTPFNTNFP
Without GANFold131653313
Fold231743322
Fold331743322
Fold431743313
Fold531653313
With GANFold131833313
Fold231743313
Fold331923322
Fold431923313
Fold531923322
Table 21

Performance results of DenseNet121 on internal dataset

KFoldAccuracyPrecisionRecall (Sensitivity)SpecificityF1-score
Without GANFold10.98780.99060.98440.99100.9875
Fold20.99080.99370.98750.99400.9906
Fold30.99080.99370.98750.99400.9906
Fold40.98930.99060.98750.99100.9891
Fold50.98780.99060.98440.99100.9875
Overall0.98930.99190.98630.99220.9891
With GANFold10.99080.99070.99070.99100.9907
Fold20.98930.99060.98750.99100.9891
Fold30.99390.99380.99380.99400.9938
Fold40.99240.99070.99380.99100.9922
Fold50.99390.99380.99380.99400.9938
Overall0.99210.99190.99190.99220.9919
Table 22

Confusion values of DenseNet169 on internal dataset

KFoldTPFNTNFP
Without GANFold131743340
Fold231833322
Fold331743331
Fold431743313
Fold531653313
With GANFold131743340
Fold231923322
Fold331833331
Fold431833331
Fold531743322
Table 23

Performance results of DenseNet169 on internal dataset

KFoldAccuracyPrecisionRecall (Sensitivity)SpecificityF1-score
Without GANFold10.99391.00000.98751.00000.9937
Fold20.99240.99380.99070.99400.9922
Fold30.99240.99690.98750.99700.9922
Fold40.98930.99060.98750.99100.9891
Fold50.98780.99060.98440.99100.9875
Overall0.99110.99440.98750.99460.9909
With GANFold10.99391.0000.98751.0000.9937
Fold20.99390.99380.99380.99400.9938
Fold30.99390.99690.99070.99700.9938
Fold40.99390.99690.99070.99700.9938
Fold50.99080.99370.98750.99400.9906
Overall0.99330.99620.99000.99640.9931
Table 24

Confusion values of VGG16 on external dataset

KFoldTPFNTNFP
Without GANFold1339660584671108
Fold2332367883661179
Fold3329670584431102
Fold4341258983161229
Fold5330569684271118
With GANFold135554468703842
Fold235814208640905
Fold334995028650895
Fold435884138629916
Fold535364658729916
Table 25

Performance results of VGG16 on external dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Withput GANFold10.87380.75400.84880.88430.7986
Fold20.86290.73810.83050.87650.7816
Fold30.86660.74940.82380.88450.7849
Fold40.86580.73520.85280.87120.7896
Fold50.86610.74720.82600.0.7847
Overall0.86700.74480.83640.87990.7879
With GANFold10.90490.80850.88850.91180.8466
Fold20.90220.79830.89500.90520.8439
Fold30.89690.79630.87450.90620.8336
Fold40.90190.79660.89680.90400.8437
Fold50.90540.81250.88380.91450.8466
Overall0.90230.80240.88770.90830.8429
Table 26

Confusion values VGG19 on external dataset

KFoldTPFNTNFP
Without GANFold1348951283481197
Fold2350649585021042
Fold335015008581964
Fold4350549683921153
Fold5352048184981047
With GANFold135994028658887
Fold236123898669876
Fold336273748721824
Fold436533488749796
Fold536093928642903
Table 27

Performance results of VGG19 on external dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.87380.74460.87200.87460.8033
Fold20.88650.77090.87630.89080.8202
Fold30.89190.78410.87500.89900.8271
Fold40.87830.75250.87600.87920.8096
Fold50.88720.77070.87980.89030.8217
Overall0.88360.76460.87580.88680.8164
With GANFold10.90480.80230.89950.90710.8401
Fold20.90660.80480.90280.90820.8510
Fold30.91160.81490.90650.91370.8583
Fold40.91550.82110.91300.91660.8646
Fold50.90440.79990.90200.90540.8479
Overall0.90860.80860.90480.91020.8540
Table 28

Confusion values of Xception on external dataset

KFoldTPFNTNFP
Without GANFold137572448575970
Fold237862158522912
Fold337732288684861
Fold437252768616929
Fold537612408631914
With GANFold138281738989556
Fold238361659033512
Fold338291729008537
Fold438531489026513
Fold538451568993552
Table 29

Performance results of Xception on external dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.91040.79480.93900.89840.8609
Fold20.91680.80590.94630.90450.8704
Fold30.91960.81420.94300.90940.8739
Fold40.91100.80040.93100.90270.8608
Fold50.91480.80450.94000.90420.8670
Overall0.91450.80390.93990.90390.8666
With GANFold10.94620.87320.95680.94170.9131
Fold20.95000.88220.95880.94640.9189
Fold30.94770.87700.95700.94370.9153
Fold40.95120.88250.96300.94620.9210
Fold50.94770.87450.96100.94220.9157
Overall0.94860.87790.95930.94400.9168
Table 30

Confusion values of ResNet50 on external dataset

KFoldTPFNTNFP
Without GANFold135574448639906
Fold235784238600945
Fold336113908570975
Fold435074948578967
Fold535644378601944
With GANFold136513509026519
Fold236123899009536
Fold336233788925620
Fold436023998953592
Fold536393628966579
Table 31

Performance results of ResNet50 on external dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.90030.79700.88900.90510.9404
Fold20.89900.79110.89430.90100.8395
Fold30.89920.78740.90250.89790.8410
Fold40.89210.78390.87650.89870.8276
Fold50.89810.79060.89080.90110.8377
Overall0.89780.79000.89060.90070.8373
With GANFold10.93580.87550.91250.94560.8936
Fold20.93170.87080.90280.94380.8865
Fold30.92630.85390.90550.93500.8789
Fold40.92680.85880.90030.93800.8791
Fold50.93050.86270.90950.93930.8855
Overall0.93030.86440.90610.94040.8847
Table 32

Confusion values of ResNet50v2 on external dataset

KFoldTPFNTNFP
Without GANFold1354845385081037
Fold2358441785231022
Fold335594428581964
Fold4354046184591086
Fold5357842384971048
With GANFold136263758822723
Fold236563428797748
Fold336123898756789
Fold436423598787758
Fold536193828771774
Table 33

Performance results of ResNet50v2 on external dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.89000.77380.88680.89140.8265
Fold20.89380.77810.89580.89290.8328
Fold30.89620.78690.88950.89900.8351
Fold40.88580.76520.88480.88620.8207
Fold50.89140.77350.89430.89020.8295
Overall0.89140.77350.89020.89190.8289
With GANFold10.91890.83380.90630.92430.8685
Fold20.91950.83020.91450.92160.8703
Fold30.91300.82070.90280.91730.8598
Fold40.91750.82770.91030.92060.8670
Fold50.91470.82380.90450.91890.8623
Overall0.91670.82720.90770.92050.8656
Table 34

Confusion values of Inceptionv3 on external dataset

KFoldTPFNTNFP
Without GANFold136014008691854
Fold235334688699846
Fold335704318697848
Fold435984038651894
Fold536133888716829
With GANFold138002019093452
Fold237692329102443
Fold337932089096449
Fold437762259083462
Fold537122859106439
Table 35

Performance results of Inceptionv3 on external dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.90740.80830.90000.91050.8517
Fold20.90300.80680.88300.91140.8432
Fold30.90560.80610.89230.91120.8481
Fold40.90430.80100.89930.90630.8473
Fold50.91020.81340.90300.91310.8559
Overall0.90610.80750.89550.91050.8492
With GANFold10.95180.89370.94980.95260.9209
Fold20.95020.89480.94200.95360.9178
Fold30.95150.89420.94800.95300.9203
Fold40.94930.89100.94380.95160.9166
Fold50.94930.89420.92780.95400.9107
Overall0.94940.89360.94230.95300.9173
Table 36

Confusion values of InceptionResNetv2 on external dataset

KFoldTPFNTNFP
Without GANFold134145878589956
Fold235054968572973
Fold336693328656889
Fold435454568596949
Fold536213808674871
With GANFold137132889017528
Fold237092928935610
Fold337252768952593
Fold437322699046599
Fold537003018953592
Table 37

Performance results of InceptionResNetv2 on external dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.88610.78120.85330.89980.8157
Fold20.89160.78270.87600.89810.8267
Fold30.90990.80500.91700.90690.8573
Fold40.89630.78880.88600.90060.8346
Fold50.90760.80610.90500.90870.8527
Overall0.89830.79280.88750.90280.8374
With GANFold10.93980.87550.92800.94470.9010
Fold20.93340.85880.92700.93610.8916
Fold30.93580.86270.93100.93790.8955
Fold40.94330.88210.93280.94770.9067
Fold50.93410.86210.92480.93800.8923
Overall0.93730.86820.92870.94090.8974
Table 38

Confusion values of DenseNet121 on external dataset

KFoldTPFNTNFP
Without GANFold135444578649896
Fold235334688643902
Fold335874148588957
Fold435814208658887
Fold535734288600945
With GANFold137003018767579
Fold237122898983562
Fold336773249002543
Fold436593428958587
Fold536623398997548
Table 39

Performance results of DenseNet121 on external dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.90010.79820.88580.90610.8397
Fold20.89890.79660.88300.90550.8376
Fold30.89880.78940.89650.89970.8396
Fold40.90350.80150.89500.90710.8457
Fold50.89860.79080.89300.90100.8388
Overall0.90000.79530.89070.90390.8403
With GANFold10.93410.86490.92480.93810.8938
Fold20.93720.86850.92780.94110.8972
Fold30.93600.87130.91900.94310.8945
Fold40.93140.86180.91450.93850.8874
Fold50.93450.86980.91530.94260.8920
Overall0.93460.86730.92030.94070.8930
Table 40

Confusion values of DenseNet169 on external dataset

KFoldTPFNTNFP
Without GANFold136673348621924
Fold236503518656886
Fold335854168628917
Fold435424598603942
Fold535554468666879
With GANFold137042978951594
Fold237102918973572
Fold336533488953592
Fold436443578956589
Fold536063958969576
Table 41

Performance results of DenseNet169 without GAN on external dataset

KFoldAccuracyPrecisionRecallSpecificityF1-score
Without GANFold10.90710.79870.91650.90320.8536
Fold20.90850.80410.91230.90690.8548
Fold30.90160.79630.89600.90390.8432
Fold40.89660.78990.88530.90130.8349
Fold50.90220.80180.88850.90790.8429
Overall0.90320.79820.89970.90460.8459
With GANFold10.93420.86180.92580.93780.8926
Fold20.93630.86640.92730.94010.8958
Fold30.93060.86050.91300.93800.8860
Fold40.93020.86090.91080.93830.8851
Fold50.92830.86230.90130.93970.8813
Overall0.93190.86240.91560.93880.8882
  15 in total

1.  Deep learning and medical diagnosis.

Authors:  Yuliang Liu; Guohua Liu; Quan Zhang
Journal:  Lancet       Date:  2019-11-09       Impact factor: 79.321

Review 2.  Review of Artificial Intelligence Techniques in Imaging Data Acquisition, Segmentation, and Diagnosis for COVID-19.

Authors:  Feng Shi; Jun Wang; Jun Shi; Ziyan Wu; Qian Wang; Zhenyu Tang; Kelei He; Yinghuan Shi; Dinggang Shen
Journal:  IEEE Rev Biomed Eng       Date:  2021-01-22

Review 3.  Applications of artificial intelligence in battling against covid-19: A literature review.

Authors:  Mohammad-H Tayarani N
Journal:  Chaos Solitons Fractals       Date:  2020-10-03       Impact factor: 5.944

4.  ai-corona: Radiologist-assistant deep learning framework for COVID-19 diagnosis in chest CT scans.

Authors:  Mehdi Yousefzadeh; Parsa Esfahanian; Seyed Mohammad Sadegh Movahed; Saeid Gorgin; Dara Rahmati; Atefeh Abedini; Seyed Alireza Nadji; Sara Haseli; Mehrdad Bakhshayesh Karam; Arda Kiani; Meisam Hoseinyazdi; Jafar Roshandel; Reza Lashgari
Journal:  PLoS One       Date:  2021-05-07       Impact factor: 3.240

5.  Text Data Augmentation for Deep Learning.

Authors:  Connor Shorten; Taghi M Khoshgoftaar; Borko Furht
Journal:  J Big Data       Date:  2021-07-19

6.  Deep learning-based multi-view fusion model for screening 2019 novel coronavirus pneumonia: A multicentre study.

Authors:  Xiangjun Wu; Hui Hui; Meng Niu; Liang Li; Li Wang; Bingxi He; Xin Yang; Li Li; Hongjun Li; Jie Tian; Yunfei Zha
Journal:  Eur J Radiol       Date:  2020-05-05       Impact factor: 3.528

7.  Data augmentation using generative adversarial networks (CycleGAN) to improve generalizability in CT segmentation tasks.

Authors:  Veit Sandfort; Ke Yan; Perry J Pickhardt; Ronald M Summers
Journal:  Sci Rep       Date:  2019-11-15       Impact factor: 4.379

8.  Application of deep learning technique to manage COVID-19 in routine clinical practice using CT images: Results of 10 convolutional neural networks.

Authors:  Ali Abbasian Ardakani; Alireza Rajabzadeh Kanafi; U Rajendra Acharya; Nazanin Khadem; Afshin Mohammadi
Journal:  Comput Biol Med       Date:  2020-04-30       Impact factor: 4.589

9.  Deep learning-based model for detecting 2019 novel coronavirus pneumonia on high-resolution computed tomography.

Authors:  Jun Chen; Lianlian Wu; Jun Zhang; Liang Zhang; Dexin Gong; Yilin Zhao; Qiuxiang Chen; Shulan Huang; Ming Yang; Xiao Yang; Shan Hu; Yonggui Wang; Xiao Hu; Biqing Zheng; Kuo Zhang; Huiling Wu; Zehua Dong; Youming Xu; Yijie Zhu; Xi Chen; Mengjiao Zhang; Lilei Yu; Fan Cheng; Honggang Yu
Journal:  Sci Rep       Date:  2020-11-05       Impact factor: 4.379

10.  Open resource of clinical data from patients with pneumonia for the prediction of COVID-19 outcomes via deep learning.

Authors:  Wanshan Ning; Shijun Lei; Jingjing Yang; Yukun Cao; Peiran Jiang; Qianqian Yang; Jiao Zhang; Xiaobei Wang; Fenghua Chen; Zhi Geng; Liang Xiong; Hongmei Zhou; Yaping Guo; Yulan Zeng; Heshui Shi; Lin Wang; Yu Xue; Zheng Wang
Journal:  Nat Biomed Eng       Date:  2020-11-18       Impact factor: 25.671

View more
  4 in total

Review 1.  Combating COVID-19 Using Generative Adversarial Networks and Artificial Intelligence for Medical Images: Scoping Review.

Authors:  Hazrat Ali; Zubair Shah
Journal:  JMIR Med Inform       Date:  2022-06-29

Review 2.  Application of machine learning in CT images and X-rays of COVID-19 pneumonia.

Authors:  Fengjun Zhang
Journal:  Medicine (Baltimore)       Date:  2021-09-10       Impact factor: 1.817

3.  An Improved COVID-19 Detection using GAN-Based Data Augmentation and Novel QuNet-Based Classification.

Authors:  Usman Asghar; Muhammad Arif; Khurram Ejaz; Dragos Vicoveanu; Diana Izdrui; Oana Geman
Journal:  Biomed Res Int       Date:  2022-02-26       Impact factor: 3.411

4.  A privacy-aware method for COVID-19 detection in chest CT images using lightweight deep conventional neural network and blockchain.

Authors:  Arash Heidari; Shiva Toumaj; Nima Jafari Navimipour; Mehmet Unal
Journal:  Comput Biol Med       Date:  2022-03-28       Impact factor: 6.698

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.