Literature DB >> 34335736

A comparative study of multiple neural network for detection of COVID-19 on chest X-ray.

Anis Shazia1, Tan Zi Xuan1, Joon Huang Chuah2, Juliana Usman1, Pengjiang Qian3, Khin Wee Lai1.   

Abstract

Coronavirus disease of 2019 or COVID-19 is a rapidly spreading viral infection that has affected millions all over the world. With its rapid spread and increasing numbers, it is becoming overwhelming for the healthcare workers to rapidly diagnose the condition and contain it from spreading. Hence it has become a necessity to automate the diagnostic procedure. This will improve the work efficiency as well as keep the healthcare workers safe from getting exposed to the virus. Medical image analysis is one of the rising research areas that can tackle this issue with higher accuracy. This paper conducts a comparative study of the use of the recent deep learning models (VGG16, VGG19, DenseNet121, Inception-ResNet-V2, InceptionV3, Resnet50, and Xception) to deal with the detection and classification of coronavirus pneumonia from pneumonia cases. This study uses 7165 chest X-ray images of COVID-19 (1536) and pneumonia (5629) patients. Confusion metrics and performance metrics were used to analyze each model. Results show DenseNet121 (99.48% of accuracy) showed better performance when compared with the other models in this study.
© The Author(s) 2021.

Entities:  

Keywords:  Artificial neural networks; COVID-19; Classification; Deep learning; DenseNet121; Multi-task learning; Transfer learning

Year:  2021        PMID: 34335736      PMCID: PMC8314263          DOI: 10.1186/s13634-021-00755-1

Source DB:  PubMed          Journal:  EURASIP J Adv Signal Process        ISSN: 1687-6172


Introduction

The novel coronavirus of 2019, or simply known as the COVID-19, affects the respiratory tracts and the lungs leading to severe cases of pneumonia. The usual symptoms include fever, dry hack cough, body ache, and loss of taste or smell. In extreme cases, the patient may experience shortness of breath and multiple organ failure and may lead to fatality (https://www.worldometers.info/coronavirus/). While the world pharmaceutical companies are trying to develop vaccination to prevent the spread of this pandemic, the current medical practice to control the spread of COVID-19 is focused on early detection and isolation of the patient. The current gold standard for COVID-19 detection is the real-time reverse transcription-polymerase chain reaction (RT-PCR), where the short sequences of DNA or RNA are reproduced or amplified and analyzed [1]. Fang et al. [2] reported that the RT-PCR testing has a low sensitivity of 71% while Williams et al. [3] reported that the sensitivity of a single RT-PCR test in hospitalized patients is 82.2%. Perceiving the limitations of RT-PCR, there is a need for cross-verification examination by using radiological images. Chest radiography, particularly chest X-ray, is one of the most frequently performed diagnostic examinations even in underdeveloped areas. Radiographic scanning was proposed to detect pathological effects of COVID-19 by examining chest radiological images of lungs of patients [4]. Several studies have shown that changes in chest radiography images such as X-ray and CT scan were noticed even before the appearance of clinical features of COVID-19 [5]. Interpretation of chest X-ray (CXR) and CT scans have widely been done by radiologists to find some visual indicators for COVID-19 infection as an alternative method for rapid screening of infected patients. For early-stage COVID-19 on CXR, peripheral ground-glass opacities are observed which progresses to consolidations at later stages [6, 7]. Since the studies have shown that the abnormalities caused by COVID-19 are visible in chest X-rays, these abnormalities, especially the opacities, are further used to detect COVID-19. Radiological COVID-19 detection also experiences challenges due to its similar nature and appearance with viral pneumonia radiographs. It requires medical experts to identify the specific radiographic markers to distinguish between the two conditions. With an enormous number of COVID-19 cases suspected daily, it is difficult to assign enough time and resources to individual radiographs. This discrepancy between the available experts and the need of the human expertise has promoted automation and machine learning to fill this much-needed gap [8]. Over the last year, scientists and researchers are unitedly working to automate the detection methods and provide intelligent machines that can easily distinguish infectious COVID-19 cases from other similar appearing cases. This study is conducted to explore these state-of-the-art techniques that have shown promising result and compare it with the same parameters and datasets to identify the best DL model for COVID-19 detection.

Related work

Jain et al. [9] implemented ResNet-101 in the classification of COVID-19 and viral pneumonia, achieving an accuracy of 97.78%. Che Azemin et al. [10] used pretrained ResNet-101 to detect COVID-19 in CXR with an accuracy of 71.9% as their training dataset was based on airspace opacity instead of confirmed COVID-19 cases. Ismael et al. [11] also used ResNet-50 architecture but only for feature extraction. The extracted features were classified using an SVM classifier with the Linear kernel function and produced high accuracy of 94.7%. Makris et al. [12] fine-tuned several CNN models and compared their performances in classifying COVID-19, pneumonia, and normal images. VGG16 turned out to have the best performance with an overall accuracy of 95.88% in their study. Abbas et al. [13] proposed a new method to classify COVID-19, SARS, and normal CXR which is called DeTraC (stands for Decompose, Transfer, and Compose). It is done by adding a class decomposition layer to the pretrained models that can partition each image class into sub-classes, but assemble back during prediction. By using VGG19 with the DeTraC approach, the model has achieved a classification accuracy of 93.1%. Asif et al. [7] trained InceptionV3 using transfer learning techniques to distinguish COVID-19 from viral pneumonia and normal CXR and obtained an accuracy of 98%. Inspired by DarkNet architecture, Ozturk et al. [14] developed a deep learning network named DarkCovidNet for automated COVID-19 diagnosis. The model achieved an accuracy of 98.08% for binary (COVID-19 and normal) and 87.02% for multiclass (COVID-19, pneumonia, and normal) classification. Shelke et al. [15] worked in the segregation of COVID-19 and normal pneumonia using DenseNet-161 and achieved an accuracy of 98.9%. Minaee et al. [16] fine-tuned 4 pretrained networks (ResNet18, ResNey50, SqueezeNet, and DenseNet-121) and compared their performance. Different cut-off thresholds for probability score were experimented in this study. SqueezNet turned out to be the best model with a sensitivity of 98% and a specificity of 92.9%. Das et al. [17] have developed a new model with a weighted average ensembling method; the model comprises of three pre-trained CNN models—DenseNet201, Resnet50V2, and InceptionV3. This approach has achieved an accuracy of 95.7% and a sensitivity of 98% in the classification of positive and negative COVID-19 cases. Ridhi et al. [18] proposed a new method to classify COVID-19, pneumonia, and normal CXR by using stacked of DenseNet and GoogleNet as feature extractor, and then the features were classified by the ensemble of XGB, RF, and SVM classifiers. The classification accuracy obtained in this study is 91.7%. Gupta et al. [19] proposed an integrated stacked deep convolution network called InstaCovNet-19 which makes use of InceptionV3, NASNet, Xception, MobileNetV2, and ResNet101. The proposed model achieved an accuracy of 99.53% in binary (COVID-19 vs non-COVID-19) classification and an accuracy of 99.08% in 3-class (COVID-19, pneumonia, normal) classification. A 22-layer CNN architecture was proposed by Hussain et al. [20] which achieved a classification accuracy of 99.1%, 94.2%, and 91.2% for binary, 3-class, and 4-class classification, respectively. Canayaz et al. [21] developed a model called MH-COVIDNet that used VGG19 as a feature extractor and BPSO meta-heuristic algorithm (MH algorithm) for feature selection. This approach obtained a classification accuracy of 99.38%. Khuzani et al. [22] performed feature extraction using different techniques such as Texture, FFT, Wavelet, GLCM, and GLDM. In the study, a multilayer network was created with 2 hidden layers of 128 and 16 neurons and a final classifier. The 3-class classification (COVID-19, pneumonia, and normal) has achieved an accuracy of 94%. From the above researches, it is observed that identification of the novel coronavirus on radiological images using deep learning techniques has the potential to reduce the pressure on radiologists. However, with various researchers using different deep learning methods, it is unclear which model provides the best result. Therefore, this study compares various deep learning models that have given impressive results in COVID-19 identification. In this study, we have fine-tuned existing models (VGG16, VGG19, DenseNet121, Inception-ResNet-V2, InceptionV3, Resnet50, and Xception) based on our classification requirements. These models have shown remarkable results in pneumonia detection [23-25] and have also been showing promising results with COVID-19 [11, 26, 27] classification. Hence, in this study, we have compared them based on the same data and variables to determine the best model to distinguish COVID-19 X-ray from pneumonia. The models have been trained and tested on COVID-19 and pneumonia CXR images from multiple datasets to avoid any biases. The models are then compared based on their performance metrics and computational time taken. The results are carefully analyzed, and the best model is chosen for this binary classification.

Materials and methods

Dataset

Due to the limitation of publicly available COVID-19 data, we have complied with multiple databases for this study. All images collected for pneumonia and COVID-19 are from publicly available datasets. Table 1 tabulates the various databases and the number of images adopted from them; similar images were eliminated. A total of 1536 COVID-19 and 5629 pneumonia images were used for training, validation, and testing of the models. The images collected from these databases were of various dimensions, which was resized to 224 × 224 pixels.
Table 1

List of databases used in this study

DatasetCOVID-19Pneumonia
1CoronaHack-Chest X-Ray-Dataset [28]194118
2Covid_Data_GradientCresent [29]69158
3Covid-19 Radiography Database [30]11431345
4Covid-chestxray-dataset [31]142
5Figure1-COVID-chestxray-dataset [32]35
6Actualmed-COVID-chestxray-dataset [33]58
7COVID-19-x-ray-10000-images [34]70
8ChestXRay2017 [35]8
Total15365629
List of databases used in this study From the total samples of COVID-19, 10% of samples was randomly selected for testing. The remaining sample was split into 80% for training and 20% for validation. Similarly, a balanced dataset was obtained by randomly selecting a similar number of samples for training and validation of pneumonia and splitting them 80% for training and 20% for validation. The remaining samples were used for testing. Table 2 tabulates the total images used in each class used for training, validation, and testing. The training and validation tests were balanced to obtain a better result and to avoid overfitting to the majority class that is the pneumonia cases. A balanced training set has been observed to give the highest accuracy regardless of the instances in the test dataset [36]. Also, the models were exposed to images from various databases to avoid any biases towards a database. Also, the imbalance between the two test sets was done to imitate a real-life environment where a number of cases are not balanced and are not from one particular source.
Table 2

Data for training, validation, and testing

COVID-19Pneumonia
Training11011104
Validation276276
Testing1564249
Data for training, validation, and testing

Transfer learning approach

There are two types of transfer learning in the context of deep learning, which are feature extraction and fine-tuning. In the feature extraction technique, a pretrained model on some standard dataset such as ImageNet is used, but the top layer, which is used for classification purpose, will be removed. Then on top of the pretrained model, it trains a new classifier to perform classification. The pretrained model without the top classifier is treated as an arbitrary feature extractor in order to extract useful features from the new dataset. In the second approach which is fine-tuning, the pretrained model weights are treated as the initial values for the new training, and they are updated and adjusted in the training process. In this case, the weights are fine-tuned from generic feature maps to specific features associated with the new dataset. The goal of fine-tuning is about adapting the generic features to a given task rather than overwriting the generic learning. For this study, a transfer learning approach was adopted and pre-trained weights from ImageNet were used to compensate for the small training data set. With transfer learning, the models were prevented from overfitting due to the small data set. In this study, we fine-tuned the last layer of seven state-of-the-art deep learning models—VGG16, VGG19, DenseNet12, Inception-ResNet-V2, InceptionV3, ResNet50, and Xception—while using the pre-trained model as a feature extractor. To fine-tune these models for binary classification, the last set of layers which consists of fully-connected layers along with softmax activation function were replaced with a flatten layer, which converts the data from the previous layer to a giant 1-dimensional tensor. A dropout of 0.5 was added for regularization, and lastly, a dense layer was added which applied softmax activation on previous layers and produce two outputs of probability for “COVID-19” and “pneumonia” classes. The next section will briefly discuss the architecture of these models and how they are used for this binary classification.

VGG16

The input of VGG16 is of fixed size 224 × 224 RGB image. It consists of 16 layers which include 13 convolutional layers and 3 fully connected layers, including max-pooling to reduce the volume size and softmax classifier following the last fully connected layer. For this study, the last fully connected layer along with softmax activation is replaced with our designed classifier as shown in Fig. 1.
Fig. 1

VGG16 architecture designed for binary classification

VGG16 architecture designed for binary classification

VGG19

The input of VGG19 is of fixed size 224 × 224 RGB image. It consists of 19 layers which include 16 convolutional layers and 3 fully connected layers, including max-pooling to reduce the volume size and softmax classifier following the last fully connected layer. For this study, the last fully connected layer along with softmax activation is replaced with our designed classifier as shown in Fig. 2.
Fig. 2

VGG19 architecture designed for binary classification

VGG19 architecture designed for binary classification

DenseNet121

The input of DenseNet121 is of fixed size 224 × 224 RGB image. DenseNet121 consists of 121 layers with parameters of more than 8 million. It is divided into DenseBlocks where the dimensions of the feature maps are the same within the block but the number of filters is different. The layers between the blocks are called transition layers and they apply batch normalization for down-sampling. For this study, the last fully connected layer along with softmax activation is replaced with our designed classifier as shown in Fig. 3.
Fig. 3

DenseNet121 architecture designed for binary classification

DenseNet121 architecture designed for binary classification

Inception-ResNet-V2

The basic building block of Inception-ResNet-V2 is called Residual Inception Block. A 1 × 1 convolution filter expansion layer is used after each block to scale up the filter bank dimensionality before the addition to match the depth of the input. This architecture uses batch normalization only on top of the traditional layers. Inception-ResNet-V2 is 164 layers deep and has an image input size of 299 × 299. The Residual Inception Block incorporates multiple-sized convolutional filters with residual connections. With the use of residual connections, this architecture prevents the problem of degradation due to deep networks and reduces the duration of training. Figure 4 explains our fine-tuned model of Inception-ResNet-V2 for COVID-19 and pneumonia classification.
Fig. 4

Inception-ResNet-V2 architecture designed for binary classification

Inception-ResNet-V2 architecture designed for binary classification

InceptionV3

InceptionV3 is made up of 484 layers consisting of 11 inception modules. It has an image input size of 299 × 299. Each module consists of convolution filters, pooling layers, and ReLu activation function. Without downgrading the network efficiency, InceptionV3 reduces the number of parameters by factorizing convolutions. InceptionV3 also proposed novel downsizing to reduce the number of features. Figure 5 explains our fine-tuned model of InceptionV3 for COVID-19 and pneumonia classification.
Fig. 5

InceptionV3 architecture designed for binary classification

InceptionV3 architecture designed for binary classification

ResNet50

ResNet50 is a variant of ResNet or Residual Network. It consists of 48 convolutional layers, 1 MaxPool, and 1 average pool layer. Each convolution block has 3 convolution layers, and there are also 3 convolution layers in each identification block. ResNet-50 has more than 23 million parameters which can be trained. Figure 6 explains our fine-tuned model of ResNet50 for COVID-19 and pneumonia classification.
Fig. 6

ResNet50 architecture designed for binary classification

ResNet50 architecture designed for binary classification

Xception

Xception was proposed by Chollet in 2016, the creator of the Keras library. It is an adaption of the Inception architectures in which the Inception modules are replaced with depth-wise separable convolutions. Xception outperformed the traditional InceptionV3 with higher Top-1 and Top-5 accuracy on ImageNet dataset. The number of parameters of Xception is roughly the same as InceptionV3 (around 23 million). Figure 7 explains our fine-tuned model of Xception for COVID-19 and pneumonia classification.
Fig. 7

Xception architecture designed for binary classification

Xception architecture designed for binary classification

Model training

For this study, all deep learning models—VGG16, VGG19, DenseNet121, Inception-ResNet-V2, InceptionV3, Resnet50, and Xception—were trained on 12 GB NVIDIA Tesla K80 GPU. All the images of the dataset were resized to 224 × 224 pixels. For the algorithm development and implementation of CNN, the deep learning library – TensorFlow 2.4 with Keras API was used. The model was trained using the categorical cross-entropy loss function to measure the performance of the model from the ground truth probabilities. The categorical cross-entropy loss function is defined as: where M indicates the class and y and p indicates the ground truth and predicted probabilities for individual images. We then minimized the loss function and improved the efficacy using Adam optimizer with a learning rate of 0.001. We implemented an early stopping technique based on validation performance to overcome the issue of overfit or underfit model. Validation loss was used as a performance measure to terminate the training when no improvement in performance was observed in 20 consecutive epochs.

Performance metrics

After the models finished training, they were tested on test set to evaluate the model accuracy. The models were tested on 156 COVID-19 images and 4249 pneumonia images. To evaluate the performance of the models, the metrics adopted include overall classification accuracy, recall (also known as sensitivity), precision, and F1-score. The metrics are defined as follow: where TP, TN, FP, and FN stand for true positive, true negative, false positive, and false negative. In this study, if the COVID-19 image is correctly classified, it is counted as TP, while if incorrectly classified as pneumonia, it is counted as FN. On the other hand, if a pneumonia image is classified correctly, it is counted as TN and the incorrectly classified as COVID-19 is FP. A confusion matrix was plotted to depict the number of correctly classified images, and a classification report was generated using the scikit-learn metrics function.

Experimental results and discussions

The accuracy and loss values in training and validation process are listed in Table 3 and shown in Figs. 8, 9, 10, 11, 12, 13, and 14 for each fine-tuned model. When comparing the number of epochs taken by each model to reach the minimum validation loss, it is observed that InceptionV3, ResNet50, and Xception reached a minimum loss at just 3, 4, and 4 epochs, respectively. With few epochs, they are able to achieve validation accuracy of 99% and above. This indicates that these models are able to learn the distinctive features between COVID-19 and pneumonia very quickly. However, when loss and accuracy are taken into consideration, it is observed that the training accuracy is highest for DenseNet121 and ResNet50; however, the DenseNet121 has the lowest training loss. For the validation set, VGG16, VGG19, DenseNet121, and Inception-ResNet-V2 have higher accuracy; however, DenseNet121 has the lowest validation loss. Hence, from this data, it can be summarized that the DenseNet121 model exhibits higher training and validation performance among the seven models.
Table 3

Accuracy and loss during training and validation

Best epochTraining lossTraining accuracyValidation lossValidation accuracy
VGG16210.01000.99900.0000e+001.0000
VGG19220.01220.99940.0000e+001.0000
DenseNet121206.1363e−091.00007.2103e−071.0000
Inception-ResNet-V290.00220.99930.00111.0000
InceptionV330.04720.99430.05540.9963
ResNet5041.1188e−041.00000.06330.9982
Xception40.04300.99510.03250.9926
Fig. 8

VGG16 accuracy loss graph

Fig. 9

VGG19 accuracy loss graph

Fig. 10

DenseNet121 accuracy loss graph

Fig. 11

Inception-ResNet-V2 accuracy loss graph

Fig. 12

InceptionV3 accuracy loss graph

Fig. 13

ResNet50 accuracy loss graph

Fig. 14

Xception accuracy loss graph

Accuracy and loss during training and validation VGG16 accuracy loss graph VGG19 accuracy loss graph DenseNet121 accuracy loss graph Inception-ResNet-V2 accuracy loss graph InceptionV3 accuracy loss graph ResNet50 accuracy loss graph Xception accuracy loss graph The confusion matrix displays numbers of images identified correctly and incorrectly by the model. The confusion matrix was generated for both the validation dataset and the test dataset. The validation dataset comprised of 276 COVID-19 and 276 pneumonia images whereas the test dataset comprised of 157 COVID-19 and 4250 pneumonia images. Table 4 below summarized the confusion matrix for all the seven models. It can be observed that though multiple models performed well during the validation, DenseNet121 has the lowest false positive and false negative, indicating that the DenseNet121 model, as shown in Fig. 15, made the least number of errors while predicting the image was COVID-19 or pneumonia.
Table 4

Confusion matrix

ModelValidationTesting
TPTNFPFNTPTNFPFN
VGG16276276001574210400
VGG19276276001574214360
DenseNet121276276001564228221
Inception-ResNet-V2276276001564172781
InceptionV3276274201554206442
Resnet50275276011564221291
Xception274274221554179712
Fig. 15

DenseNet121 confusion matrix

Confusion matrix DenseNet121 confusion matrix This study even compared these pre-trained models based on the accuracy, precision, recall, and F1 score as tabulated in Table 5. It is observed that DenseNet121 gave good classification performance with an accuracy of 99.48%, followed by ResNet50 with 99.32% accuracy. Table 5 also compares the computational times taken by each model for training and testing. It is seen that the InceptionV3 takes the least time (11 min 50 s) for training; however, it is slow during testing (16 min 14 s), whereas DenseNet121 was slower during training (20 min), but it was the fastest during testing (15 min 36 s) with the highest accuracy.
Table 5

Performance metrics

ModelAccuracy (%)Precision (%)Recall (%)F1 (%)Training time (s)Testing time (s)
VGG1699.0999.2899.0999.1413531012
VGG1999.1899.3499.1899.221344946
DenseNet12199.4899.5499.4899.491200936
Inception-ResNet-V298.2198.7998.2198.388991083
InceptionV398.9699.1798.9699.02710974
ResNet5099.3299.4299.3299.357831056
Xception98.3498.8398.3498.49750865
Performance metrics From the above result, we recommend DenseNet121 (99.48% accuracy, 99.54% precision, 99.48% recall, and 99.49% of F1 score) for classification of COVID-19 from pneumonia cases on chest X-ray, further comparing our fine-tuned DenseNet121 with the studies recently published that also performed binary classification, particularly COVID-19 and pneumonia images. Shelke et al. [15] used a deeper network, DenseNet-161, but the accuracy obtained is lower, which might be due to the lower number of training images. Compared with other works that worked on the binary classification of CXR images, our model has the second highest accuracy (Table 6). The highest binary classification accuracy is obtained by Gupta et al. [19] using their proposed network called InstaCovNet-19.
Table 6

Comparison with other related work that performed binary classification on CXR images

Author (year)CNN architecturesImage classificationTotal datasetAccuracy
Nayak et al. [37]ResNet-34COVID-19 and normal

COVID-19: 775

Normal: 775

98.33%
Ismael and Şengür [11]ResNet50 feature + SVMCOVID-19 and normal

COVID-19: 180

Normal: 200

94.7%
Ozturk et al. [14]DarkCovidNetCOVID-19 and no-findings

COVID-19: 177

No-findings: 500

98.08%
Amit Kumar et al. [38]Ensembling method: three pretrained CNN models—DenseNet201, Resnet50V2, and Inceptionv3COVID-19 positive and COVID-19 negative

COVID-19 +ve: 538

COVID-19 −ve: 468

95.7%
Gupta et al. [19]InstaCovNet-19 (integrated stacking of InceptionV3, NASnet, Xception, MobileNetV2 and ResNet101)COVID-19 and non-COVID

COVID-19: 361

Normal: 365

99.53%
Hussain et al. [20]CoroDetCOVID-19 and normal

COVID-19: 2843

Normal: 1108

99.1%
Shelke et al. [15]DenseNet-161COVID-19 and pneumonia

COVID-19: 500

Pneumonia: 500

98.9%
Jain et al. [9]ResNet-101COVID-19 and viral pneumonia

COVID-19: 440

Viral pneumonia: 480

97.78%
This studyDenseNet-121COVID-19 and pneumonia

COVID-19: 1536

Pneumonia: 5629

99.48%
Comparison with other related work that performed binary classification on CXR images COVID-19: 775 Normal: 775 COVID-19: 180 Normal: 200 COVID-19: 177 No-findings: 500 COVID-19 +ve: 538 COVID-19 −ve: 468 COVID-19: 361 Normal: 365 COVID-19: 2843 Normal: 1108 COVID-19: 500 Pneumonia: 500 COVID-19: 440 Viral pneumonia: 480 COVID-19: 1536 Pneumonia: 5629

Conclusion

Deep learning algorithm can aid healthcare workers in detecting COVID-19 with minimal processing of chest X-ray images. In this study, 2-class datasets were created which included COVID-19 and pneumonia images obtained from open sources. Several state-of-the-art pretrained neural networks that include ResNet50, DenseNet121, InceptionV3, VGG16, VGG19, Inception-ResNet-V2, and Xception were experimented using transfer learning technique. The best model turned out to be DenseNet-121 which accomplished an accuracy of 99.48%, followed by ResNet50 with a classification accuracy of 99.32%. This study summarizes that the detection models built using CNNs with transfer learning technique are able to perform good binary classification tasks on COVID-19 and pneumonia images. COVID-19 and viral pneumonia CXR images contain similar features which are challenging for the radiologist to interpret. However, the CNN model can easily learn the features in just a few epochs of training and classify the images correctly. The high accuracies obtained suggest that the deep learning models could find something distinctive in the CXR images and that makes the deep networks capable of distinguishing the images correctly. These trained models can effectively reduce the workload of medical practitioners and increase the accuracy and efficiency of COVID-19 diagnosis.
  20 in total

1.  Sensitivity of Chest CT for COVID-19: Comparison to RT-PCR.

Authors:  Yicheng Fang; Huangqi Zhang; Jicheng Xie; Minjie Lin; Lingjun Ying; Peipei Pang; Wenbin Ji
Journal:  Radiology       Date:  2020-02-19       Impact factor: 11.105

2.  Automated detection of COVID-19 cases using deep neural networks with X-ray images.

Authors:  Tulin Ozturk; Muhammed Talo; Eylul Azra Yildirim; Ulas Baran Baloglu; Ozal Yildirim; U Rajendra Acharya
Journal:  Comput Biol Med       Date:  2020-04-28       Impact factor: 4.589

3.  A familial cluster of pneumonia associated with the 2019 novel coronavirus indicating person-to-person transmission: a study of a family cluster.

Authors:  Jasper Fuk-Woo Chan; Shuofeng Yuan; Kin-Hang Kok; Kelvin Kai-Wang To; Hin Chu; Jin Yang; Fanfan Xing; Jieling Liu; Cyril Chik-Yan Yip; Rosana Wing-Shan Poon; Hoi-Wah Tsoi; Simon Kam-Fai Lo; Kwok-Hung Chan; Vincent Kwok-Man Poon; Wan-Mui Chan; Jonathan Daniel Ip; Jian-Piao Cai; Vincent Chi-Chung Cheng; Honglin Chen; Christopher Kim-Ming Hui; Kwok-Yung Yuen
Journal:  Lancet       Date:  2020-01-24       Impact factor: 79.321

4.  Detection of 2019 novel coronavirus (2019-nCoV) by real-time RT-PCR.

Authors:  Victor M Corman; Olfert Landt; Marco Kaiser; Richard Molenkamp; Adam Meijer; Daniel Kw Chu; Tobias Bleicker; Sebastian Brünink; Julia Schneider; Marie Luisa Schmidt; Daphne Gjc Mulders; Bart L Haagmans; Bas van der Veer; Sharon van den Brink; Lisa Wijsman; Gabriel Goderski; Jean-Louis Romette; Joanna Ellis; Maria Zambon; Malik Peiris; Herman Goossens; Chantal Reusken; Marion Pg Koopmans; Christian Drosten
Journal:  Euro Surveill       Date:  2020-01

5.  Application of deep learning techniques for detection of COVID-19 cases using chest X-ray images: A comprehensive study.

Authors:  Soumya Ranjan Nayak; Deepak Ranjan Nayak; Utkarsh Sinha; Vaibhav Arora; Ram Bilas Pachori
Journal:  Biomed Signal Process Control       Date:  2020-11-19       Impact factor: 3.880

6.  InstaCovNet-19: A deep learning classification model for the detection of COVID-19 patients using Chest X-ray.

Authors:  Anunay Gupta; Shreyansh Gupta; Rahul Katarya
Journal:  Appl Soft Comput       Date:  2020-10-29       Impact factor: 6.725

7.  The role of balanced training and testing data sets for binary classifiers in bioinformatics.

Authors:  Qiong Wei; Roland L Dunbrack
Journal:  PLoS One       Date:  2013-07-09       Impact factor: 3.240

8.  Deep-COVID: Predicting COVID-19 from chest X-ray images using deep transfer learning.

Authors:  Shervin Minaee; Rahele Kafieh; Milan Sonka; Shakib Yazdani; Ghazaleh Jamalipour Soufi
Journal:  Med Image Anal       Date:  2020-07-21       Impact factor: 8.545

9.  A deep learning approach to detect Covid-19 coronavirus with X-Ray images.

Authors:  Govardhan Jain; Deepti Mittal; Daksh Thakur; Madhup K Mittal
Journal:  Biocybern Biomed Eng       Date:  2020-09-07       Impact factor: 4.314

View more
  2 in total

Review 1.  Medical image processing and COVID-19: A literature review and bibliometric analysis.

Authors:  Rabab Ali Abumalloh; Mehrbakhsh Nilashi; Muhammed Yousoof Ismail; Ashwaq Alhargan; Abdullah Alghamdi; Ahmed Omar Alzahrani; Linah Saraireh; Reem Osman; Shahla Asadi
Journal:  J Infect Public Health       Date:  2021-11-17       Impact factor: 3.718

2.  COVID-19 classification using chest X-ray images: A framework of CNN-LSTM and improved max value moth flame optimization.

Authors:  Ameer Hamza; Muhammad Attique Khan; Shui-Hua Wang; Abdullah Alqahtani; Shtwai Alsubai; Adel Binbusayyis; Hany S Hussein; Thomas Markus Martinetz; Hammam Alshazly
Journal:  Front Public Health       Date:  2022-08-30
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.