Literature DB >> 34219953

Deep convolution neural networks to differentiate between COVID-19 and other pulmonary abnormalities on chest radiographs: Evaluation using internal and external datasets.

Yongwon Cho1, Sung Ho Hwang1, Yu-Whan Oh1, Byung-Joo Ham2, Min Ju Kim1, Beom Jin Park1.   

Abstract

We aimed to evaluate the performance of convolutional neural networks (CNNs) in the classification of coronavirus disease 2019 (COVID-19) disease using normal, pneumonia, and COVID-19 chest radiographs (CXRs). First, we collected 9194 CXRs from open datasets and 58 from the Korea University Anam Hospital (KUAH). The number of normal, pneumonia, and COVID-19 CXRs were 4580, 3884, and 730, respectively. The CXRs obtained from the open dataset were randomly assigned to the training, tuning, and test sets in a 70:10:20 ratio. For external validation, the KUAH (20 normal, 20 pneumonia, and 18 COVID-19) dataset, verified by radiologists using computed tomography, was used. Subsequently, transfer learning was conducted using DenseNet169, InceptionResNetV2, and Xception to identify COVID-19 using open datasets (internal) and the KUAH dataset (external) with histogram matching. Gradient-weighted class activation mapping was used for the visualization of abnormal patterns in CXRs. The average AUC and accuracy of the multiscale and mixed-COVID-19Net using three CNNs over five folds were (0.99 ± 0.01 and 92.94% ± 0.45%), (0.99 ± 0.01 and 93.12% ± 0.23%), and (0.99 ± 0.01 and 93.57% ± 0.29%), respectively, using the open datasets (internal). Furthermore, these values were (0.75 and 74.14%), (0.72 and 68.97%), and (0.77 and 68.97%), respectively, for the best model among the fivefold cross-validation with the KUAH dataset (external) using domain adaptation. The various state-of-the-art models trained on open datasets show satisfactory performance for clinical interpretation. Furthermore, the domain adaptation for external datasets was found to be important for detecting COVID-19 as well as other diseases.
© 2021 Wiley Periodicals LLC.

Entities:  

Keywords:  COVID‐19; chest radiography; computer‐aided diagnosis (CAD); deep learning; lung diseases

Year:  2021        PMID: 34219953      PMCID: PMC8239912          DOI: 10.1002/ima.22595

Source DB:  PubMed          Journal:  Int J Imaging Syst Technol        ISSN: 0899-9457            Impact factor:   2.177


area under the curve computer‐aided diagnosis convolutional neural network coronavirus disease 2019 computed tomography chest radiograph deep learning Korea University Anam Hospital receiver operating characteristic curve

INTRODUCTION

The coronavirus disease 2019 (COVID‐19) pandemic, characterized by the severe acute respiratory syndrome and caused by coronavirus 2, has continued almost unabated since the virus was first detected in December 2019. There have been around 27 000 000 confirmed cases and 875 000 confirmed deaths worldwide at the time this report was written (September 8, 2020). Early detection of COVID‐19 is crucial to prevent infection of healthy people owing to the highly contagious nature of the virus. Currently, reverse transcriptase‐polymerase chain reaction (RT‐PCR), which can detect SARS‐CoV‐2 RNA in suspected patients, is the primary detection method for COVID‐19. However, this method is time‐consuming (from 3 h to more than 48 h) and involves complicated manual processes. Chest radiography imaging, including computed tomography (CT) or X‐ray imaging is typically used for pneumonia diagnosis. CT screening for the initial diagnosis of COVID‐19 has been found to be superior to RT‐PCR testing and can even confirm COVID‐19 infection after a negative or weakly positive diagnosis performed using RT‐PCR testing. Recently, CT imaging has been employed in various studies for the diagnosis of COVID‐19. , However, with the spread of COVID‐19, the routine application of CT places a substantial strain on the radiology department. Therefore, the role of chest radiography has become increasingly crucial with regard to the first diagnostic imaging parameters for screening patients with nonspecific thoracic symptoms for pneumonia or COVID‐19 in general clinical practice. Generally, chest radiographs (CXRs) reflect thoracic symptoms, including bilateral, peripheral consolidation, and/or ground‐glass opacities on CT. , Wong et al. investigated the sensitivity of COVID‐19 diagnosis using CXRs. Unfortunately, CXR results have been reported to be less sensitive compared with initial RT‐PCR tests (69 vs. 91%), although 9% of negatively diagnosed cases using RT‐PCR were later diagnosed positively using CXRs. In addition, automatic methods to detect subtle abnormalities such as COVID‐19 and pneumonia can supplement clinical tools for early diagnosis under conditions such as a large number of potentially infected people and a limited number of trained radiologists. Although CXRs cannot be used to entirely substitute RT‐PCR tests, pneumonia is a clinical symptom in high‐risk patients who require hospitalization; therefore, CXRs can be used for patient triage, to determine the priority of patient treatment in order to assist overloaded healthcare systems in the midst of a global pandemic. This is especially important because the most frequently known cause of community‐acquired pneumonia is bacterial infection. Excluding these populations through triage can significantly reduce the amount of medical resources needed. Accordingly, artificial intelligence, including deep learning (DL), are potential solutions for classification of COVID‐19. It is highly challenging to collect a large volume of high‐quality curated CXRs to train DL networks in clinical environments. Recently, however, research using small datasets of COVID‐19 CXRs has been actively explored. , , , , , , , , In particular, Wang and Wong presented an open‐source network (called the COVID‐Net) to detect COVID‐19 using CXRs. This network showed good performance, with 80% sensitivity. Minaeea et al. analyzed the performance of various open‐source (shared online) DL algorithms for COVID‐19 detection using CXRs. These algorithms achieved a high sensitivity of 98% (±3%) and a specificity of approximately 90%. In this study, we aimed to evaluate the COVID‐19 diagnosis performance of a customized state‐of‐the‐art (SOTA) DL algorithms. These algorithms were developed using open datasets for training and internal validation. DL architectures, which can used for radiologically interpreting inference results, were trained on a dataset containing a limited number of normal, pneumonia, and COVID‐19 images. We provide a statistical analysis of the performance of three models, namely, DenseNet169, InceptionResNetV2, and Xception. Importantly, a Korea University Anam Hospital (KUAH) dataset was used to evaluate the COVID‐19 diagnosis performance for external validation using three models trained on multiple open datasets. We also demonstrate that COVID‐19 classification performance can be improved by using domain adaptation, such as performing histogram matching on external datasets to reduce the heterogeneities between open (internal) and KUAH (external) datasets.

MATERIALS AND METHODS

Our institutional review board approved our retrospective cohort study, and the requirement for informed consent was waived.

Datasets

CXR datasets are of two types: open or local. We employed a CXR dataset that has been used in the development of existing SOTA algorithms, COVID‐CXNet and COVID‐Net. The dataset containing COVID‐19, normal, and pneumonia images were collected from multiple public sources. The most important feature of this dataset was that it contained a total of 805 COVID‐19 images. Because COVID‐19 is a new disease, its image database is regularly updated on various public sites such as Radiopedia, SIRM, EuroRad, and Hannover Medical School dataset. In addition, a non‐COVID‐19 dataset, containing a large number of normal and pneumonia images, was collected from the RSNA Pneumonia Detection Challenge 2018. Further, this dataset contains 5000 normal and 4272 pneumonia images. This dataset includes lung opacity and various conditions such as bleeding, volume loss, pulmonary edema, and lung cancer, which also cause opacity in CXRs. Although this was a segmentation problem to find main lesions, we used classification to classify COVID‐19 and other diseases in the present study. A total of 10 077 normal, pneumonia, and COVID‐19 images from the open datasets were randomly assigned into training, tuning, and test sets in a 70:10:20 ratio for the final computer‐aided diagnosis (CAD) assessment for detecting COVID‐19 using open CXRs (Table 1). For external validation, we collected a small dataset from the KUAH. These datasets were selected depending on the availability of the corresponding chest CT images from January 2020 to October 2020, which were confirmed by expert thoracic radiologists. The number of normal, pneumonia, and COVID‐19 CXRs were 20, 20, and 18, respectively (Table 1).
TABLE 1

Number of CXR images in the training and test sets

ClassesThe severity of COVID‐19Number of images in the training set (with tuning set)Number of images in the test set
Internal dataset (multiple open datasets)Normal3780 (420)800
Pneumonia3484 (388)400
COVID‐19Early phase605 (50)125
External dataset KUAHNormal20
Pneumonia20
COVID‐19Severe phase18

Note: KUAH for external validation.

Abbreviations: CXR, chest radiograph; KUAH, Korea University Anam Hospital.

Number of CXR images in the training and test sets Note: KUAH for external validation. Abbreviations: CXR, chest radiograph; KUAH, Korea University Anam Hospital. The characteristics of the COVID‐19 images in the open and KUAH datasets are as follows: In both the cases, COVID‐19 infection was confirmed using RT‐PCR testing. The COVID‐19 patients whose images were included in the open datasets were likely in the early phase of disease progression, whereas those whose images were included in the KUAH dataset were in a critically severe phase of disease progression. Such cases are difficult to distinguish from those of conventional pneumonia, even for expert radiologists, without the aid of CT. In addition, patients with pneumonia and other diseases in the KUAH dataset were at a severe level of disease progression. As the open datasets used for training and internal validation are different from that employed for external validation (KUAH), the interpretation of the results becomes crucial and can be used to determine the way the results should be used in a real medical environment such as at the KUAH. Figure 1(A) shows 15 sample images from multiple open datasets, including five normal and pneumonia images from the RSNA pneumonia challenge (first and second rows), and five COVID‐19 images from various public sites (third row). Figure 1(B) shows 15 sample images from the KUAH dataset, including normal (first row), pneumonia (second row), and COVID‐19 (third row) images.
FIGURE 1

Examples from (A) multiple public datasets for training and internal validation, and (B) Korea University Anam Hospital (KUAH) dataset for external validation. Normal, pneumonia, and COVID‐19 images are in the first, second, and third rows, respectively, of (A) and (B)

Examples from (A) multiple public datasets for training and internal validation, and (B) Korea University Anam Hospital (KUAH) dataset for external validation. Normal, pneumonia, and COVID‐19 images are in the first, second, and third rows, respectively, of (A) and (B)

Methods

We classified the three classes, including COVID‐19 CXRs and others, using a pretrained DenseNet that employed transfer learning, InceptionResNetV2, and Xception.

DenseNet

DenseNet is configured with a dense block that comprises four BN‐ReLU‐Conv modules, as shown in Figure 2. The colored squares represent feature maps generated at different steps. The convolution layers are most commonly used for image convolution.
FIGURE 2

DenseNet architecture: It has three dense blocks and the layers between two adjacent blocks are referred to as transition layers that change the size of the feature‐maps via convolution and pooling

DenseNet architecture: It has three dense blocks and the layers between two adjacent blocks are referred to as transition layers that change the size of the feature‐maps via convolution and pooling

InceptionResNetV2

InceptionResNetV2 is formulated based on a combination of the inception structure and residual connections. In the Inception‐ResNet block, convolutional filters of various sizes are combined with residual connections. The use of residual connections not only helps us avoid the degradation problem caused by deep structures but also reduces the training time. Figure 3 shows the architecture of InceptionResNetV2.
FIGURE 3

Architecture for InceptionResNetV2: InceptionResNetV2 is a convolutional neural network that is trained on more than a million images from the ImageNet database. It has 164 deep layers and can classify images into 1000 objects. The input size is 299 × 299 for the network

Architecture for InceptionResNetV2: InceptionResNetV2 is a convolutional neural network that is trained on more than a million images from the ImageNet database. It has 164 deep layers and can classify images into 1000 objects. The input size is 299 × 299 for the network

Xception

We proposed a convolutional neural network (CNN) architecture based entirely on depthwise separable convolution layers. In effect, the mapping of cross‐channel correlations and spatial correlations in the feature maps of CNNs can be entirely decoupled. Because this hypothesis is a stronger version of the one underlying the inception architecture, we refer to this as the “Xception” architecture, which stands for “extreme inception.” A complete description of the network specifications is presented in Figure 4. The Xception architecture has 36 convolutional layers, forming the feature extraction base of the network.
FIGURE 4

Architecture for Xception: Xception is highly efficient in terms two main points: depthwise shortcuts and separable convolution between convolution blocks, as in ResNet

Architecture for Xception: Xception is highly efficient in terms two main points: depthwise shortcuts and separable convolution between convolution blocks, as in ResNet The network was trained on CXRs based on weak labels and fine‐tuned using DenseNet169, InceptionResNetV2, and Xception, all pretrained on the ImageNet dataset to classify normal, pneumonia, and COVID‐19 using images from multiple open datasets. The three CNN models were trained on more than 1 million natural images, including 1000 objects in ImageNet. We used transfer learning after the network was trained on a substantially large amount of labeled data (i.e., pretrained on the National Institutes of Health [NIH] dataset). These deep CNN techniques enable the learning of generic image features by employing other domain datasets without the need for training the network from scratch. The pretrained networks serve as feature extractors for generic image features, and the last two layers are fully connected for classification. We trained the models that were pretrained on the NIH dataset with CXRs (Table 1; internal dataset) and only fine‐tuned the last layer of each deep CNN model. For evaluation, we used the internal and external datasets corresponding to each model. The image size was 512 × 512 pixels. A careful redesign of the workflows with regard to preprocessing, deep CNNs, and computing hardware settings was undertaken. We used geometric augmentation, including zoom, rotation, and shift, on the edges of the images. These augmentations helped alleviate scanner‐specific biases and improved the robustness of the CNNs against additional sources of variability, unrelated to the radiological classes in Python 3.6. In addition, we devised multiscale and mixed (MM)‐COVID‐19Net to train the DL models, as shown in Figure 5. This method was devised based on previous research. It differs from Reference 27 in that the patch images were randomly extracted from the whole images and were rescaled to multiple sizes and various zoom factors. This was intended to reflect the manner in which a radiologist might determine various disease patterns on specific CXRs. Multiscale patch images and whole images were regularly used to train each deep CNN on the CXR data. The input images were resized to 512 × 512 pixels and converted into NumPy arrays. These datasets were loaded on a GPU server with Ubuntu 18.04, CUDA 10.2, four 24 GB Titan RTX and Quadro graphics cards, and cuDNN 9.1 (NVIDIA Corporation) with the Keras (TensorFlow) framework. We used the Adam or RAdam optimizer with an initial learning rate of 0.001 for the classification of COVID‐19 images. The cross‐entropy cost function in binary classification (1.1) is expressed as follows: where f and y denote the inferred probability and corresponding desired output, respectively.
FIGURE 5

Overall architecture of the multiscale and mixed (MM)‐COVID‐19Net for the classification of COVID‐19

Overall architecture of the multiscale and mixed (MM)‐COVID‐19Net for the classification of COVID‐19 Tuning errors in the selection of optimized models were minimized by running the backpropagation algorithm over 25 training epochs with a batch size of eight. After training, we conducted an ablation study on the classification of COVID‐19, normal, and pneumonia images using a test dataset comprising images from the open datasets (internal dataset) and the KUAH dataset (external dataset) to confirm the correctness of the results of the feature pyramid and dense connection information. The external validation dataset (KUAH) was preprocessed using histogram matching to enhance the detection rate of COVID‐19 by adjusting the external dataset with regard to differences such as texture and intensity from multiple open datasets (internal). This method converts images to a histogram that can be specified arbitrarily. A template image from the internal datasets must be selected for histogram matching, as shown in Figure 6.
FIGURE 6

Illustration of the histogram matching process in our datasets. T = template image

Illustration of the histogram matching process in our datasets. T = template image Furthermore, this method examines the headers used in all digital imaging and communications in medicine (DICOM) images to determine the intensity range (L, W) because CXRs are not standardized with respect to intensity. The DICOM header includes information regarding the window width (0028, 1051) and window center (0028, 1050). The method uses this information to calculate the highest pixel value (1.2) and the lowest pixel value (1.3) in the image. where P is the lowest pixel value, P is the highest pixel value, and P  and P are the center and width intensities of the input pixel, respectively. Because the histograms of the images are distributed in various forms, we investigated the relationship between the histogram and the intensity value of all the CXRs to select a template image for histogram matching. First, we calculated the mean intensity of the selected corresponding images using (1.4). The standard template image is a uniformly distributed histogram, as shown in Figure 6. The performance of CAD for COVID‐19 classification was evaluated for each model with and without the new preprocessing method using histogram matching. In addition, we proposed gradient‐weighted class activation mapping (Grad‐CAM) to visualize the feature map associated with each class for any convolutional layer of the networks using the gradient generated via backpropagation.

Statistical analysis

We evaluated the diagnostic performance of the CNN models for the classification of COVID‐19, normal, and pneumonia images using fivefold cross‐validation analysis. We defined terms that form the confusion matrix as follows: true positive (TP) is the number of labels correctly classified as positive by the algorithms, true negative (TN) is the number of labels correctly classified as negative by the algorithms, false positive (FP) is the number of labels incorrectly classified as positive by the algorithms, and false negative (FN) is the number of labels incorrectly classified as negative by the algorithms. Multiple classifications based on CXRs were assessed in terms of recall, precision, F1‐score, and accuracy as follows: These values were calculated using the scikit‐learn Python library. Accuracy is the ratio of the number of correctly classified test samples to the total number of test samples. For the multiclass case, the area under the curve (AUC) (COVID‐19‐vs‐all) was calculated using the pROC (1.17.0.1) R package.

RESULTS

Comparison of MM‐COVID‐19Net with three CNNs

To predict COVID‐19, MM‐COVID‐19Net combined with three widely used CNNs, namely, DenseNet169, InceptionResNetV2, and Xception, as the backbone network, as shown in Figure 5, was trained and validated with fivefold cross‐validation with internal validation on multiple open datasets containing normal and abnormal classes (COVID‐19 and pneumonia, respectively). For additional external validation, we used the KUAH dataset (normal, 20 images; COVID‐19, 18 images; and pneumonia, 20 images). The models calculated a probability value for each image to be detected as either a COVID‐19 image or an image corresponding to the other classes. We can also use a binary label indicating whether or not an image corresponds to COVID‐19 to calculate the recall, precision, F1‐score, and accuracy. The best performing algorithm was chosen from the trained models using fivefold cross‐validation. Considering the KUAH dataset (external) is different from the multiple open datasets (internal), as shown in Table 1, an independent external validation is more important than internal validation to accurately evaluate applicability in medical environments; therefore, CAD was conducted on each model, with and without histogram matching for domain adaptation, and the results were evaluated. The results are presented in Table 2 and Figure 7. A model for COVID‐19 detection should have high sensitivity, and thus, the best model was selected based on the recall value. The COVID‐19 scores for Xception were as follows: 97.50 ± 0.92 for recall, 98.39 ± 1.28 for precision, 97.50 ± 0.92 for F1‐score, 0.99 ± 0.01 for AUC, and 93.57 ± 0.29% for accuracy, averaged over all folds; Xception exhibited its best performance in the fifth‐fold with scores of 98.4, 98.4, 98.4, 0.9997, and 93.74% for recall, precision, F1‐score, AUC, and accuracy, respectively. The COVID‐19 scores for DenseNet169 were 96.00 ± 1.13, 98.53 ± 1.46, 97.25 ± 1.11, 0.99 ± 0.01, and 92.94% ± 0.45% for recall, precision, F1‐score, AUC, and accuracy, respectively, averaged over all folds, with its best performance in the first fold, where the scores were 97.6, 100, 98.78, 0.9998, and 93.36% for recall, precision, F1‐score, AUC, and accuracy, respectively. The COVID‐19 scores for InceptionResNetV2 were 96.48 ± 0.91, 97.73 ± 0.66, 97.10 ± 0.53, 0.99 ± 0.01, and 93.12 ± 0.23% for recall, precision, F1‐score, AUC, and accuracy, respectively, averaged over all folds, with its best performance in the fifth‐fold, wherein the scores were 96.80, 97.58, 97.19, 0.9997, and 93.28% for recall, precision, F1‐score, AUC, and accuracy, respectively. Although the difference was not significant, the Xception architecture exhibited the best detection performance among the three CNNs in terms of all metrics (*Xception‐fifth‐fold: p‐values: 0.91, 0.83, 0.14, and 0.14 among DenseNet169 and InceptionResNetV2 in MM‐COVID‐19Net, and References 9 and 10).
TABLE 2

Results of the fivefold cross validation for classification (normal, pneumonia, and COVID‐19 images) on the internal dataset) with (MM)‐COVID‐19Net‐backbone network: DenseNet169, Reference [9], and Reference [10]. Avg = average; SD = standard deviation on multiple open dataset (internal validation)

FoldLabelRecallPrecisionF1‐scoreAUC for multiclassAccuracy
MM‐COVID‐19Net‐backbone network: Xception 1 Normal95.6393.9894.800.976593.43
Pneumonia87.7590.9389.310.9704
COVID‐19 97.6 97.6 97.6 0.9998
2 Normal95.2593.9694.600.974693.13
Pneumonia88.2589.3788.800.9675
COVID‐19 97.54 1.0 97.54 0.9997
3 Normal95.6394.5695.090.978493.81
Pneumonia89.9290.6189.920.9723
COVID‐19 97.98 99.18 97.98 0.9997
4 Normal95.6394.9195.270.976193.74
Pneumonia89.9490.4089.950.9688
COVID‐19 95.97 96.75 95.97 0.9995
5*, # Normal95.2594.6694.950.976493.74
Pneumonia89.2590.3889.810.9707
COVID‐19 98.4 98.4 98.4 0.9997
Avg ± SD Normal95.48 ± 0.2194.41 ± 0.4294.94 ± 0.260.97 ± 0.01 93.57 ± 0.29
Pneumonia89.02 ± 0.9990.34 ± 0.5889.56 ± 0.500.97 ± 0.01
COVID‐19 97.50 ± 0.92 98.39 ± 1.28 97.50 ± 0.92 0.99 ± 0.01
MM‐COVID‐19Net‐backbone network: DenseNet 1691*Normal94.7594.5194.630.979493.36
Pneumonia89.2589.0289.140.9717
COVID‐19 97.6 1.0 98.78 0.9998
2 Normal95.5093.6394.550.976692.75
Pneumonia86.5089.6488.040.9701
COVID‐19 95.20 96.75 95.97 0.9996
3 Normal94.5094.3894.400.980392.91
Pneumonia89.0087.9088.440.9747
COVID‐19 95.20 1.0 97.54 0.9992
4 Normal95.3894.4394.900.980693.36
Pneumonia88.2589.5988.920.9748
COVID‐19 96.80 98.37 97.58 0.9995
5 Normal94.2593.9094.070.979292.30
Pneumonia87.5087.5087.500.9708
COVID‐19 95.20 97.54 96.36 0.9994
Avg ± SD Normal94.88 ± 0.5594.17 ± 0.3894.51 ± 0.310.97 ± 0.01 92.94 ± 0.45
Pneumonia88.10 ± 1.1388.73 ± 0.9888.41 ± 0.660.97 ± 0.01
COVID‐19 96.00 ± 1.13 98.53 ± 1.46 97.25 ± 1.11 0.99 ± 0.01
MM‐COVID‐19Net‐backbone network: InceptionResNetV2 1 Normal95.6393.7594.680.976693.28
Pneumonia87.5090.9089.170.9713
COVID‐19 96.80 97.58 97.19 0.9997
2 Normal95.0093.8394.410.977092.91
Pneumonia87.7589.5488.630.9701
COVID‐19 96.00 97.56 96.77 0.9996
3 Normal95.1394.3094.710.979493.21
Pneumonia88.2589.8289.030.9751
COVID‐19 96.80 96.80 96.80 0.9996
4 Normal95.7593.1994.450.979292.83
Pneumonia86.2590.3188.240.9741
COVID‐19 95.20 98.34 96.75 0.9997
5* Normal95.8893.6594.750.975293.36
Pneumonia87.0091.1089.000.9693
COVID‐19 97.60 98.39 97.99 0.9997
Avg ± SD Normal95.48 ± 0.3993.74 ± 0.4094.6 ± 0.160.98 ± 0.01 93.12 ± 0.23
Pneumonia87.35 ± 0.7690.33 ± 0.6788.81 ± 0.380.97 ± 0.02
COVID‐19 96.48 ± 0.91 97.73 ± 0.66 97.1 ± 0.53 0.99 ± 0.01
Reference [9] COVID‐CXNet 1* Normal95.1390.3892.690.964290.19
Pneumonia77.7589.6383.270.9533
COVID‐19 98.4 90.44 94.25 0.9949
2 Normal93.8890.3792.090.956089.43
Pneumonia81.0084.8182.860.9438
COVID‐19 88.00 98.21 92.83 0.9928
3 Normal97.6388.7592.980.965389.81
Pneumonia72.7592.9781.630.9546
COVID‐19 94.4 89.39 91.83 0.9952
4 Normal90.3891.8791.150.954784.68
Pneumonia69.0089.6177.970.9320
COVID‐19 98.4 53.48 62.30 0.9913
5 Normal88.8893.9291.320.954286.72
Pneumonia89.0073.2580.360.9439
COVID‐19 65.60 1.0 79.22 0.9882
Avg ± SD Normal93.18 ± 3.5591.06 ± 1.9492.05 ± 0.810.96 ± 0.01 88.17 ± 2.38
Pneumonia77.79 ± 7.7286.51 ± 8.0481.22 ± 2.140.95 ± 0.02
COVID‐19 88.96 ± 13.73 86.30 ± 18.93 84.09 ± 13.58 0.99 ± 0.01
Reference [10] COVID‐Net 1 Normal88.6394.1691.300.959589.13
Pneumonia88.5078.8483.390.9478
COVID‐19 94.4 95.93 95.16 0.9940
2 Normal98.5088.0492.970.966890.64
Pneumonia74.096.1083.610.9585
COVID‐19 93.6 95.90 94.73 0.9980
3 Normal91.2593.9592.580.971190.19
Pneumonia88.5081.9485.090.9650
COVID‐19 88.88 95.69 92.21 0.9959
4* Normal96.6390.1993.300.968791.01
Pneumonia79.091.3384.720.9618
COVID‐19 95.90 95.90 94.74 0.9981
5 Normal90.3894.8892.570.973590.49
Pneumonia90.5080.8085.380.9661
COVID‐19 91.20 99.13 99.13 0.9974
Avg ± SD Normal93.08 ± 4.2692.24 ± 2.9892.54 ± 0.760.97 ± 0.0290.29 ± 0.71
Pneumonia84.10 ± 7.2085.80 ± 7.5084.43 ± 0.890.96 ± 0.03
COVID‐19 92.80 ± 2.77 96.51 ± 1.47 95.19 ± 2.49 0.99 ± 0.01

Note: In the table, # denotes p‐values: 0.91, 0.83, 0.14, and 0.14 among DenseNet169 and InceptionResNetV2 in MM‐COVID‐19Net and References [9] and [10] and * denotes the best performance for each model, respectively.

Abbreviations: AUC, area under the curve; MM, multiscale and mixed.

FIGURE 7

Confusion matrix on multiple open datasets (internal). The first, middle, and last columns present the results of the multiscale and mixed (MM)‐COVID19‐Net (Xception, DenseNet169, InceptionResnetV2), COVID‐CXNet, and COVID‐Net, respectively (0: normal; 1: pneumonia; 2: COVID‐19)

Results of the fivefold cross validation for classification (normal, pneumonia, and COVID‐19 images) on the internal dataset) with (MM)‐COVID‐19Net‐backbone network: DenseNet169, Reference [9], and Reference [10]. Avg = average; SD = standard deviation on multiple open dataset (internal validation) Note: In the table, # denotes p‐values: 0.91, 0.83, 0.14, and 0.14 among DenseNet169 and InceptionResNetV2 in MM‐COVID‐19Net and References [9] and [10] and * denotes the best performance for each model, respectively. Abbreviations: AUC, area under the curve; MM, multiscale and mixed. Confusion matrix on multiple open datasets (internal). The first, middle, and last columns present the results of the multiscale and mixed (MM)‐COVID19‐Net (Xception, DenseNet169, InceptionResnetV2), COVID‐CXNet, and COVID‐Net, respectively (0: normal; 1: pneumonia; 2: COVID‐19) For the internal dataset, the results of each model were determined using Grad‐CAM for normal and abnormal (pneumonia and COVID‐19) images after the application of DenseNet169, Xception, and InceptionResNetV2, as shown in Figure 8.
FIGURE 8

Gradient‐weighted class activation mapping (Grad‐CAM) results for each model (normal, pneumonia, and COVID‐19) on chest radiographs (CXRs). The first column shows the CXRs; the remainder, from the second to the fourth columns, are the corresponding heatmaps for each model. (The red color indicates a normal tissue in the normal images and abnormalities in the abnormal images)

Gradient‐weighted class activation mapping (Grad‐CAM) results for each model (normal, pneumonia, and COVID‐19) on chest radiographs (CXRs). The first column shows the CXRs; the remainder, from the second to the fourth columns, are the corresponding heatmaps for each model. (The red color indicates a normal tissue in the normal images and abnormalities in the abnormal images) In addition, we compared the performance of our proposed method (histogram matching, domain adaptation) with the original preprocessing method on the KUAH dataset (external) for COVID‐19 and other conditions, as presented in Table 1. The results are presented in Table 3 and Figure 9. The accuracies of the best models in Xception (fifth‐fold model), DenseNet169 (first‐fold model), and InceptionResNetV2 (fifth‐fold model) with histogram matching were 68.97, 74.14, and 68.97%, respectively, and those without histogram matching were 62.07, 68.97, and 60.34%, respectively. The recalls corresponding to COVID‐19 prediction for each algorithm with histogram matching were 50.00, 55.56, and 50.00%, respectively, and those without histogram matching were 33.33, 38.89, and 38.89%, respectively (p = 0.0123; Table 3 and Figure 9).
TABLE 3

Results of the best model among the fivefold cross‐validation for classification (normal, pneumonia, and COVID‐19) on the KUAH (external)

With histogram matchingLabelRecallPrecisionF1‐scoreAUC for multiclassAccuracy
MM‐COVID‐19Net‐backbone network: Xception Yes*Normal90.0085.7187.800.9540 68.97
Pneumonia65.0065.0065.000.7763
COVID‐19 50.00 52.94 51.40 0.7153
NoNormal95.0070.3780.850.919762.07
Pneumonia55.0061.1157.890.7855
COVID‐19 33.3346.1538.70 0.6431
MM‐COVID‐19Net‐backbone network: DenseNet169 Yes*Normal95.0079.1186.360.9303 74.14
Pneumonia70.0073.6871.950.8105
COVID‐19 55.56 66.67 60.60 0.7472
NoNormal95.0076.0084.440.910568.97
Pneumonia70.0066.6768.290.8569
COVID‐19 38.8958.3346.670.6819
MM‐COVID‐19Net‐backbone network: InceptionResNetV2 Yes*Normal95.0073.0882.610.9500 68.97
Pneumonia60.0070.5964.860.8319
COVID‐19 50.00 60.00 54.54 0.7653
NoNormal90.0064.2975.000.929060.34
Pneumonia50.0071.4258.820.8658
COVID‐19 38.8943.7541.720.6556

Note: The performance on the external dataset was evaluated with the best model for each architecture (with histogram matching). *p‐value: 0.0123 without histogram matching.

Abbreviations: AUC, area under the curve; KUAH, Korea University Anam Hospital; MM,, multiscale and mixed.

FIGURE 9

Confusion matrix for the Korea University Anam Hospital (KUAH) dataset (external) in Table 1. The first, middle, and last columns show the results of multiscale and mixed (MM)‐COVID19‐Net (Xception, DenseNet169, and InceptionResnetV2). (A) Results of the classification without histogram matching. (B) Results of the classification with histogram matching

Results of the best model among the fivefold cross‐validation for classification (normal, pneumonia, and COVID‐19) on the KUAH (external) Note: The performance on the external dataset was evaluated with the best model for each architecture (with histogram matching). *p‐value: 0.0123 without histogram matching. Abbreviations: AUC, area under the curve; KUAH, Korea University Anam Hospital; MM,, multiscale and mixed. Confusion matrix for the Korea University Anam Hospital (KUAH) dataset (external) in Table 1. The first, middle, and last columns show the results of multiscale and mixed (MM)‐COVID19‐Net (Xception, DenseNet169, and InceptionResnetV2). (A) Results of the classification without histogram matching. (B) Results of the classification with histogram matching Furthermore, we determined the classification results and localization using Grad‐CAM for normal and abnormal (pneumonia and COVID‐19) images after the application of DenseNet169, Xception, and InceptionResNetV2 with and without histogram matching, as shown in Figures 10 and 11. The CAM results for all trained classes were visualized independently, and the CAMs of the regions of interest were extracted individually.
FIGURE 10

Confusion matrix for the Korea University Anam Hospital (KUAH) dataset (external) in Table 1. The first, middle, and last columns show the results of multiscale and mixed (MM)‐COVID19‐Net (backbone: DenseNet169) COVID‐CXNet, and COVID‐Net, respectively. (0: normal; 1: pneumonia; 2: COVID‐19)

FIGURE 11

Gradient‐weighted class activation mapping (Grad‐CAM) for COVID‐19 (ground truth) on chest radiographs. (A) Negative results of each model corresponding to normal or pneumonia images without histogram matching. (B) Positive results of each model indicating COVID‐19 with histogram matching

Confusion matrix for the Korea University Anam Hospital (KUAH) dataset (external) in Table 1. The first, middle, and last columns show the results of multiscale and mixed (MM)‐COVID19‐Net (backbone: DenseNet169) COVID‐CXNet, and COVID‐Net, respectively. (0: normal; 1: pneumonia; 2: COVID‐19) Gradient‐weighted class activation mapping (Grad‐CAM) for COVID‐19 (ground truth) on chest radiographs. (A) Negative results of each model corresponding to normal or pneumonia images without histogram matching. (B) Positive results of each model indicating COVID‐19 with histogram matching

Comparison of the MM‐COVID‐19Net (the best model), COVID‐CXNet, and COVID‐Net

We trained COVID‐CXNet and COVID‐Net on CXRs in the same way that we trained the MM‐COVID‐19Net. These networks were compared with our network through statistical analysis on multiple open datasets (internal) and the KUAH dataset (external). Among the trained models with fivefold cross‐validation of the backbone network, Xception with MM‐COVID‐19Net, the best performing network, was selected using the fifth‐fold model. The first and fourth folds were selected for COVID‐CXNet and COVID‐Net, respectively, as presented in Table 2. Using the internal dataset for COVID‐19, the scores for our algorithm were as follows: 98.40% for recall, 98.40% for precision, 98.40% for F1‐score, 0.9997 for AUC, and 93.74% for accuracy. The corresponding values for COVID‐CXNet were as follows: 98.40% for recall, 90.44% for precision, 94.25% for F1‐score, 0.9949 for AUC, and 90.19% for accuracy; and for COVID‐Net these values were as follows: 95.90% for recall, 95.90% for precision, 94.74% for F1‐score, 0.9981 for AUC, and 91.01% for accuracy. Using the external dataset, among the trained models with fivefold cross‐validation of the backbone network, the scores of the MM‐COVID‐19Net using DenseNet169 for our algorithm were as follows: 55.56% for recall, 66.67% for precision, 60.60% for F1‐score, 0.7472 for AUC, and 74.14% for accuracy. The corresponding values for COVID‐CXNet were as follows: 0% for recall, zero for precision, zero for F1‐score, 0.3444 for AUC, and 44.83% for accuracy (p‐value = 0.004); whereas the scores for COVID‐Net were as follows: 33.33% for recall, 85.71% for precision, 48.00% for F1‐score, 0.5918 for AUC, and 77.59% for accuracy (p‐value = 0.82).

DISCUSSION AND CONCLUSION

We analyzed the feasibility of classifying COVID‐19 images using SOTA DL algorithms trained on a dataset containing a limited number of normal, pneumonia, and COVID‐19 images obtained from open datasets to radiologically interpret the inference results. For external validation, we used the KUAH dataset, which was confirmed by expert radiologists using CT. Furthermore, we developed a new training architecture based on SOTA to reflect the manner in which a radiologist might determine various disease patterns on specific CXRs. Previous studies , used statistical analyses to assess the AUC, accuracy, sensitivity, and specificity. These analyses were conducted using open datasets; however, we aimed to determine whether CNN architectures trained using multiple open datasets (internal) performed sufficiently well for use in real clinical environments. In addition, previous studies have been conducted on the application of CAD in response to nodules and various disease patterns , , , using high‐quality CXRs and without an open dataset. We herein investigated the detection and classification of abnormalities, including thoracic disease patterns, and provided insights for the detection of COVID‐19. Our results indicate that all the models have similar COVID‐19 detection performance. Among the three CNNs, the accuracy of Xception was 93.57% ± 0.29% with 97.50% recall and 0.99 ± 0.01 AUC for COVID‐19 on the internal datasets. A comparison between our algorithm and Reference 9, which achieved an AUC of 0.99 ± 0.01 and accuracy of 88.17 ± 2.38%, and Reference 10, which achieved an AUC of 0.99 ± 0.01 and accuracy of 90.29 ± 0.71%, showed no significant difference between our algorithm and other algorithms. However, it is important to indirectly compare the characteristics of open datasets with those of datasets obtained in a specific medical environment. Although the number of images in the KUAH dataset (external) was limited, as presented in Table 1, and the results presented in Table 3 are inferior to those obtained using multiple open datasets (internal), as presented in Table 2, COVID‐19 images could be detected by the CNNs trained using the open datasets. DenseNet169 proved to be the best model, with an accuracy of 68.97% and recall of 38.89%. These results suggest that COVID‐19 patients whose images were included in the multiple open datasets (internal) are likely in the early stages of disease progression and those whose images are included in the KUAH dataset (external) are in highly severe stages of disease progression, which are difficult to distinguish from pneumonia without the aid of CT, even for expert radiologists. Therefore, we conducted histogram matching on the KUAH dataset, that is, domain adaptation, to improve the results using the external dataset. Using histogram matching, we matched the KUAH dataset (external) with template images from the training dataset (internal dataset composed of multiple open datasets). The results obtained show that the performance of the algorithm after domain adaptation, such as performing histogram matching (accuracy: 74.14%, recall of COVID‐19: 55.56%, and AUC of COVID‐19: 0.7472), was better than that without histogram matching on the KUAH dataset (Figure 9 and Table 3). In addition, we compared our algorithm with others , on the KUAH dataset (external). The results for Reference 9 (accuracy: 44.83%, recall of COVID‐19: 0%, and AUC of COVID‐19: 0.3444) and Reference 10 (accuracy: 77.59%, recall of COVID‐19: 33.33%, and AUC of COVID‐19: 0.5918) were lower than those of our algorithm, as presented in Table 4. Therefore, our algorithm can be useful when training (internal—open) and testing (external—KUAH) with different datasets.
TABLE 4

Comparison of the best models for classification (normal, pneumonia, and COVID‐19) on the KUAH dataset (external with domain adaptation for histogram matching) in MM‐COVID‐19Net‐backbone network: DenseNet169, COVID‐CXNet, and COVID‐Net

LabelRecallPrecisionF1‐scoreAUC for multiclassAccuracy
MM‐COVID‐19Net‐backbone network: DenseNet169*Normal95.0079.1186.360.930374.14
Pneumonia70.0073.6871.950.8105
COVID‐19 55.56 66.67 60.60 0.7472
Reference [9] COVID‐CXNetNormal95.0054.2969.090.875044.83
Pneumonia35.0041.1737.880.5171
COVID‐19 0 0 0 0.3444
Reference [10] COVID‐NetNormal95.0070.3780.850.965877.59
Pneumonia1.0083.3390.090.8948
COVID‐19 33.33 85.71 48.00 0.5918

Note: Performance on the external dataset was evaluated with the best model for MM‐COVID‐19Net backbone network: DenseNet169 (with histogram matching). *p‐value: 0.004 and 0.82 among and References [9] and [10].

Abbreviations: AUC, area under the curve; MM, multiscale and mixed.

Comparison of the best models for classification (normal, pneumonia, and COVID‐19) on the KUAH dataset (external with domain adaptation for histogram matching) in MM‐COVID‐19Net‐backbone network: DenseNet169, COVID‐CXNet, and COVID‐Net Note: Performance on the external dataset was evaluated with the best model for MM‐COVID‐19Net backbone network: DenseNet169 (with histogram matching). *p‐value: 0.004 and 0.82 among and References [9] and [10]. Abbreviations: AUC, area under the curve; MM, multiscale and mixed. Although the COVID‐19 classification performance was insufficient on the external dataset, it is vital for distinguishing other viral pneumonia from COVID‐19 pneumonia. Even though highly accurate detection was not achieved on the KUAH dataset (external), distinguishing the risk groups for the purpose of screening is still important. In addition, if internal or external datasets are well refined and various preprocessing techniques are used, the performance can be improved for distinguishing traditional pneumonia from COVID‐19‐induced pneumonia. Salehi et al. described COVID‐19 patterns in CXRs, including peripheral distribution, ground‐glass opacification, and bilateral involvement. The results of various DL algorithms should reflect such radiological findings and patterns. Therefore, Grad‐CAM was used to visualize the interpretation of the DL models, making them more transparent. This approach showed that the models derive information that indicates not only the presence of the disease but also, indirectly, the disease location. The visualization results presented in the first and second rows depicted in Figure 11 coincide with the opinions of radiologists, although some radiologists might question whether the last column actually represents the location of the lesion. Nevertheless, Grad‐CAM shows that the model with histogram matching is more accurate than that without histogram matching, as demonstrated in Figure 11. Our study has several limitations. First, there is a lack of well‐curated CXRs (open datasets with COVID‐19 images) that could be obtained from multiple public datasets (see Table 1). Moreover, the open COVID‐19 datasets for training were largely collected websites and online publications; thus, strict standards were not applied when collecting them. The characteristics of COVID‐19 are not diverse, this could have an impact when training the models using open datasets and testing them for external validation on real‐world clinical datasets, such as the KUAH dataset. In addition, no comparisons were made between the results of the human observers and those of the models in the detection of COVID‐19 and other diseases. Furthermore, owing to a lack of GPU memory, all the CXRs were down sampled to 512 × 512 pixels, which could decrease the clinical classification and detection validity. Finally, although our algorithm can help improve the detection of COVID‐19, we must train our algorithm using various real‐world clinical datasets. In the future, we intend to train the algorithm using various CXRs including COVID‐19 and pneumonia to improve the detection performance, particularly for COVID‐19, in clinical settings. After collecting more CXRs that include various disease patterns independently confirmed by expert radiologists in our institutions and additional centers, we intend to develop algorithms that can diagnose COVID‐19 and other diseases based on multiple CXRs. Finally, histogram matching will be used to detect lesions using datasets different from those of the training datasets (multiple open datasets). We will therefore develop a superior domain adaptation method to improve the detection of COVID‐19 based on CXRs. In addition, we will train our models on real clinical datasets (the KUAH dataset and various multicenter datasets) to develop our algorithms on CXRs and evaluate these models using multiple open datasets or others. Most importantly, the diagnosis results obtained by human radiologists should be compared with those obtained by CNN algorithms through reader tests. In conclusion, we evaluated three widely used CNNs for the classification of COVID‐19 and other diseases using a limited number of CXRs for internal (open datasets) and external validation (KUAH). These models, trained using open datasets, were validated on an external dataset (KUAH), which was representative of actual clinical environments. Although the diagnostic performances of the models using the internal dataset (open datasets) were desirable (Table 2), the performance of the same models on the external dataset (KUAH) was insufficient for routine clinical use. To overcome this problem, we proposed domain adaption, such as histogram matching, to detect or classify COVID‐19 and other diseases more accurately on external datasets. This method produced better results for dataset obtained from special hospitals. Although COVID‐19 detection using CXRs (internal or external datasets) has the limitation in that it cannot enable a direct confirmation of COVID‐19 in the same way as a diagnostic kit using an RT‐PCR test, it can still assist in identifying lung diseases caused by respiratory infections in clinical environments. Our empirical evaluation of the diagnosis of COVID‐19 can be extended to the development of CAD algorithms for COVID‐19 based on DL for use in real‐world clinical environments.

CONFLICT OF INTEREST

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflicts of interest.

AUTHOR CONTRIBUTIONS

Yongwon, Cho, Beom Jin Park, and Sung Ho Hwang: Wrote the manuscript. Yongwon Cho: Performed the experiments and prepared the Figures. Yu‐Whan Oh, Beom Jin Park, Byung‐Joo Ham, and Sung Ho Hwang: Prepared the dataset and confirmed the datasets. Min Ju Kim: Confirmed the datasets. All authors reviewed the manuscript. All authors were involved in writing the paper and approved the final submitted and published versions.
  21 in total

1.  An Ensemble of Fine-Tuned Convolutional Neural Networks for Medical Image Classification.

Authors:  Ashnil Kumar; Jinman Kim; David Lyndon; Michael Fulham; Dagan Feng
Journal:  IEEE J Biomed Health Inform       Date:  2016-12-05       Impact factor: 5.772

2.  Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks.

Authors:  Ali Narin; Ceren Kaya; Ziynet Pamuk
Journal:  Pattern Anal Appl       Date:  2021-05-09       Impact factor: 2.580

3.  COVID-19 on Chest Radiographs: A Multireader Evaluation of an Artificial Intelligence System.

Authors:  Keelin Murphy; Henk Smits; Arnoud J G Knoops; Michael B J M Korst; Tijs Samson; Ernst T Scholten; Steven Schalekamp; Cornelia M Schaefer-Prokop; Rick H H M Philipsen; Annet Meijers; Jaime Melendez; Bram van Ginneken; Matthieu Rutten
Journal:  Radiology       Date:  2020-05-08       Impact factor: 11.105

4.  Reproducibility of abnormality detection on chest radiographs using convolutional neural network in paired radiographs obtained within a short-term interval.

Authors:  Yongwon Cho; Young-Gon Kim; Sang Min Lee; Joon Beom Seo; Namkug Kim
Journal:  Sci Rep       Date:  2020-10-15       Impact factor: 4.379

5.  COVID-CXNet: Detecting COVID-19 in frontal chest X-ray images using deep learning.

Authors:  Arman Haghanifar; Mahdiyar Molahasani Majdabadi; Younhee Choi; S Deivalakshmi; Seokbum Ko
Journal:  Multimed Tools Appl       Date:  2022-04-07       Impact factor: 2.577

6.  Deep-COVID: Predicting COVID-19 from chest X-ray images using deep transfer learning.

Authors:  Shervin Minaee; Rahele Kafieh; Milan Sonka; Shakib Yazdani; Ghazaleh Jamalipour Soufi
Journal:  Med Image Anal       Date:  2020-07-21       Impact factor: 8.545

7.  Frequency and Distribution of Chest Radiographic Findings in Patients Positive for COVID-19.

Authors:  Ho Yuen Frank Wong; Hiu Yin Sonia Lam; Ambrose Ho-Tung Fong; Siu Ting Leung; Thomas Wing-Yan Chin; Christine Shing Yen Lo; Macy Mei-Sze Lui; Jonan Chun Yin Lee; Keith Wan-Hang Chiu; Tom Wai-Hin Chung; Elaine Yuen Phin Lee; Eric Yuk Fai Wan; Ivan Fan Ngai Hung; Tina Poy Wing Lam; Michael D Kuo; Ming-Yen Ng
Journal:  Radiology       Date:  2020-03-27       Impact factor: 11.105

8.  Covid-19: automatic detection from X-ray images utilizing transfer learning with convolutional neural networks.

Authors:  Ioannis D Apostolopoulos; Tzani A Mpesiana
Journal:  Phys Eng Sci Med       Date:  2020-04-03

9.  Chest CT for Typical Coronavirus Disease 2019 (COVID-19) Pneumonia: Relationship to Negative RT-PCR Testing.

Authors:  Xingzhi Xie; Zheng Zhong; Wei Zhao; Chao Zheng; Fei Wang; Jun Liu
Journal:  Radiology       Date:  2020-02-12       Impact factor: 11.105

10.  A Curriculum Learning Strategy to Enhance the Accuracy of Classification of Various Lesions in Chest-PA X-ray Screening for Pulmonary Abnormalities.

Authors:  Beomhee Park; Yongwon Cho; Gaeun Lee; Sang Min Lee; Young-Hoon Cho; Eun Sol Lee; Kyung Hee Lee; Joon Beom Seo; Namkug Kim
Journal:  Sci Rep       Date:  2019-10-25       Impact factor: 4.379

View more
  1 in total

1.  Deep convolution neural networks to differentiate between COVID-19 and other pulmonary abnormalities on chest radiographs: Evaluation using internal and external datasets.

Authors:  Yongwon Cho; Sung Ho Hwang; Yu-Whan Oh; Byung-Joo Ham; Min Ju Kim; Beom Jin Park
Journal:  Int J Imaging Syst Technol       Date:  2021-05-13       Impact factor: 2.177

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.