Literature DB >> 33037291

A comprehensive study on classification of COVID-19 on computed tomography with pretrained convolutional neural networks.

Tuan D Pham1.   

Abstract

The use of imaging data has been reported to be useful for rapid diagnosis of COVID-19. Although computed tomography (CT) scans show a variety of signs caused by the viral infection, given a large amount of images, these visual features are difficult and can take a long time to be recognized by radiologists. Artificial intelligence methods for automated classification of COVID-19 on CT scans have been found to be very promising. However, current investigation of pretrained convolutional neural networks (CNNs) for COVID-19 diagnosis using CT data is limited. This study presents an investigation on 16 pretrained CNNs for classification of COVID-19 using a large public database of CT scans collected from COVID-19 patients and non-COVID-19 subjects. The results show that, using only 6 epochs for training, the CNNs achieved very high performance on the classification task. Among the 16 CNNs, DenseNet-201, which is the deepest net, is the best in terms of accuracy, balance between sensitivity and specificity, [Formula: see text] score, and area under curve. Furthermore, the implementation of transfer learning with the direct input of whole image slices and without the use of data augmentation provided better classification rates than the use of data augmentation. Such a finding alleviates the task of data augmentation and manual extraction of regions of interest on CT images, which are adopted by current implementation of deep-learning models for COVID-19 classification.

Entities:  

Mesh:

Year:  2020        PMID: 33037291      PMCID: PMC7547710          DOI: 10.1038/s41598-020-74164-z

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

Image findings have been increasingly recognized as a useful tool for rapid diagnosis of COVID-19[1]. The use of chest computed tomography (CT) to detect COVID-19 symptoms has been reported to have high sensitivity and can predate positive tests carried out in a laboratory[2-7]. Latest articles on image analysis of COVID-19 can be further found at the European Radiology webaite[8]. Because of the potential utilization of CT data, hospitals overloaded with admissions of patients are using CT imaging to diagnose COVID-19 infection and to decide the order of treatment of infected patients. In fact, chest CT has an important role in urgent clinical assessment and decision making for treatment of COVID-19 patients who suffer from severe and worsening respiratory symptoms. In other words, CT scans can be used to assess the severity of the infected lungs as well as progress of the disease, which tremendously help medical doctors in curbing the virus. It has been suggest that examinations and reports of CT findings can be used as a basis for improving the quality of care for COVID-19 patents[9]. While CT imaging is useful for the diagnosis of COVID-19, manual reading of the scans is time-consuming and subject to human error. Therefore, the need for advanced artificial intelligence (AI)-based automated image analysis has the potential to analyze CT scans in the assessment of COVID-19. AI-based image analysis methods can provide accurate and rapid diagnosis of the disease to cope with the demand for a large number of patients[10]. For example, a manual assessment of a CT scan can take up to 15 minutes, while AI-based image analysis requires only a few seconds. Furthermore, AI can be useful in improving clinical workflow efficiency[11]. A recent baseline study on AI for automated classification of COVID-19 using the largest publicly available CT dataset was reported in[12], which will be described subsequently as the database used in this study. This work used the pretrained DenseNet for the classification task. These authors adopted transfer learning and data augmentation for training the pretrained deep-learning model with new image data. The rationale is that transfer learning can relieve the need for acquiring a large amount of training data by reusing a developed model as the starting point for training a new model with a different task. The data augmentation was performed by using a large dataset of chest X-ray images. The rationale for image data augmentation is to increase the size of the training dataset with plausible examples in order to improve the performance and ability of the deep-learning model to generalize the power of classification by getting familiar with samples of high variance. Another recent work reported on the use of ten pretrained CNNs for classifying CT scans of COVID-19 and non-COVID-19 subjects[13]. These authors reported that ResNet-101 and Xception provided the best classification results on training and testing a CT dataset consisting 106 COVID-19 patients and 86 non-COVID-19 subjects. The CNNs were trained and tested with regions of interest extracted from the CT scans that were defined by a radiologist. Other previous works on the classification of COVID-19 on CT scans were reported in[14-16]. A 3D deep-learning network was developed for the detection of COVID-19 from 4356 3D chest CT scans obtained from 3322 patients[14]. The network extracted both 2D local and 3D global features from the CT scans. This network, called COVNet, was built on the pretrained RestNet50. In[15], the pretrained Inception was modified to detect COVID-19 using extracted regions of interest on CT scans obtained from 180 cases of COVID-19 and 79 cases of SARs-COV-2. In[16], a total of 618 CT scans were used, consisting of 219 CT scans from 110 COVID-19 patients, 224 CT scans from 224 patients with Influenza-A viral pneumonia, and 175 CT scans from healthy people. Pulmonary regions of interest were extracted from the CT scans, and pretrained ResNet-18 was used for image feature extraction. Finally, the Noisy-or Bayesian function was used to classify the image regions into three types: COVID-19, Influenza-A-viral-pneumonia, and irrelevant-to-infection. However, it should be noted that the CT datasets used in the studies reported in[13-16] are not publicly available. In this study, a comprehensive investigation on 16 pretrained CNNs for classification of COVID-19 using a publicly available CT database is presented. These pretrained CNNs reflect a variety of computational complexity and accuracy based on the training and testing of the ImageNet database[17]. Findings of this investigation would facilitate the timely deployment of AI-assisted tools to hospitals and clinics in terms of ease of both data preparation and software implementation for fighting against the pandemic.

Methods

COVID-19 CT database

The COVID-19 CT database used in this study is publicly available[18], and its details are described in[12]. The database consists of 349 CT images containing clinical findings of COVID-19 from 216 patients, and 397 CT images obtained from non-COVID-19 subjects. These CT images were collected from COVID19-related papers published in medRxiv, bioRxiv, NEJM, JAMA, Lancet, and others. Figure 1 shows CT images of COVID-19 and non-COVID-19. The usefulness of this dataset has been confirmed by a senior radiologist of Tongji Hospital, Wuhan, China. The radiologist has performed diagnosis and treatment of a large number of COVID-19 patients during the virus outbreak between January and April 2020[18].
Figure 1

CT images: Rows 1 and 2: COVID-19, Rows 3 and 4: non-COVID-19 (aspect ratios of some images were rescaled to fit the figure frame).

CT images: Rows 1 and 2: COVID-19, Rows 3 and 4: non-COVID-19 (aspect ratios of some images were rescaled to fit the figure frame).

Implementation of pretrained CNNs

Sixteen pretrained CNNs were investigated in this study for the classification of whole CT images to differentiate COVID-19 from non-COVID-19. These networks are: (1) SqueezeNet, (2) GoogLeNet, (3) Inception-v3, (4) DenseNet-201, (5) MobileNet-v2, (6) ResNet-18, (7) ResNet-50, (8) ResNet-101, (9) Xception, (10) Inception-ResNet-v2, (11) ShuffleNet, (12) NasNet-Mobile, (13) NasNet-Large, (14) AlexNet, (15) VGG-16, and (16) VGG-19. These pretrained networks were trained on more than a million images from the ImageNet database[17]. The pretrained networks can classify images into 1000 object categories, such as keyboard, mouse, pencil, and many animals. As a result, these networks have learned rich features representing a wide range of images. The properties of these networks are described in Table 1. To enable the reproduction of the results reported in this study, configurations for the transfer learning are described as follows.
Table 1

Properties of 16 pre-trained CNNs.

CNNDepthSize (MB)Parameters (millions)Input image size
AlexNet822761.0227 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 227
GoogLeNet22277.0224 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 224
SqueezeNet184.61.24227 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 227
ShuffleNet506.31.4224 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 224
ResNet-18184411.7224 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 224
ResNet-50509625.6224 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 224
ResNet-10110116744.6224 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 224
Xception718522.9299 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 299
Inception-v3488923.9299 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 299
Inception-ResNet-v216420955.9299 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 299
VGG-1616515138224 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 224
VGG-1919535144224 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 224
DenseNet-2012017720.0224 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 224
MobileNet-v253133.5224 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 224
NasNet-Mobile*205.3224 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 224
NasNet-Large*36088.9331 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}× 331

*indicates NASNet-Mobile and NasNetLarge networks do not consist of a linear sequence of modules.

Properties of 16 pre-trained CNNs. *indicates NASNet-Mobile and NasNetLarge networks do not consist of a linear sequence of modules. First, the layer graph from the pretrained network was extracted. If the network was a SeriesNetwork object, such as AlexNet, VGG-16, or VGG-19, then the list of layers was converted to a layer graph. In most pretrained networks, the last layer with learnable weights is a fully connected layer. This fully connected layer was replaced with a new fully connected layer with the number of outputs equal to the number of classes in the new data set, which is 2, in this study. In some pretrained networks, such as SqueezeNet, the last learnable layer is a 1-by-1 convolutional layer instead. In this case, the convolutional layer was replaced with a new convolutional layer with the number of filters equal to the number of classes. For the option of data augmentation in this study, random reflection, translation, and scaling were carried out. Random reflection was done in the top-bottom direction, where each image was reflected vertically with probability = 0.5. The range of horizontal translation applied to the input image = [− 30, 30], where the translation distance was measured in pixels. The horizontal translation distance was selected randomly from a continuous uniform distribution within the specified range. Similarly, the interval of vertical translation applied to the input image in pixels = [− 30, 30]. The vertical translation distance was selected randomly from a continuous uniform distribution within the specified interval. The range of horizontal scaling was applied to the input image, where the horizontal scale factor was selected randomly from a continuous uniform distribution within the specified interval = [0.9, 1.1]. Similarly, the range of vertical scaling was applied to the input image, where the vertical scale factor was selected randomly from a continuous uniform distribution within the specified interval = [0.9, 1.1]. The original whole CT images were converted into RGB images and resized to fit into the input image size of each pretrained CNN. For the training options, the stochastic gradient descent with momentum optimizer was used, where the momentum value = 0.9000; gradient threshold method =  norm; minimum batch size = 10; maximum number of epochs = 6; initial learning rate = 0.0003; the learning rate remained constant throughout training; the training data were shuffled before each training epoch, and the validation data were shuffled before each network validation; and factor for regularization (weight decay) = 0.0001.

Statistical measures of classification performance

Five statistical measures used for evaluating the two-class classification performance of the pretrained CNNs are accuracy, sensitivity, specificity, score, and the area under the receiver operating characteristic (ROC) curve (AUC). Let the sensitivity (SEN) be the percentage of COVID-19 patients who are correctly identified as having the infection, and expressed aswhere TP is called true positive, denoting the number of COVID-19 patients who are correctly identified as having the infection, FN false negative, denoting the number of COVID-19 patients who are misclassified as having no infection of COVID-19, and P the total number of COVID-19 patients. Let the specificity (SPE) be the percentage of non-COVID-19 subjects who are correctly classified as having no infection of COVID-19:where TN is called true negative and denotes the number of non-COVID-19 subjects who are correctly identified as having no infection of COVID-19, FP false positive, denoting the number of non-COVID-19 subjects who are misclassified as having the infection, and N the total number of non-COVID-19 subjects. The percent accuracy (ACC) of the classification is defined asThe score is defined as the balance between precision (TP divided by TP and FP) and sensitivity: The ROC is a probability curve created by plotting the TP rate against the FP rate at various threshold settings, and the AUC represents the measure of performance of a classifier. The AUC value is within the range between 0.5 and 1, where the value = 0.5 represents the performance of a random classifier and the value = 1 indicates a perfect one. Thus, the higher the AUC is, the better the classifier performs. The AUC was calculated using the trapezoidal integration to estimate the area under the ROC curve.

Results

To compare the results with those obtained from previous reports, the dataset was randomly split into 80% for training and 20% for testing. The data splitting was repeated 5 times to obtain the average and standard deviation for each CNN. The whole CT images were used as the data input, which were resized to fit the input image size of each pretrained CNN, in both training and testing phases. The network training was performed for with and without data augmentation. Tables 2 and 3 show the classification results obtained from the 16 CNNs with and without data augmentation, respectively.
Table 2

Classification results with data augmentation.

CNN modelAccuracy (%)Sensitivity (%)Specificity (%)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$F_1$$\end{document}F1 scoreAUC
AlexNet74.50 ± 4.4070.46 ± 6.3779.05 ± 8.610.75 ± 0.040.83 ± 0.04
GoogLeNet78.97 ± 3.7075.95 ± 13.6982.38 ± 10.530.79 ± 0.060.91 ± 0.04
SqueezeNet78.52 ± 7.5691.56 ± 7.6363.81 ± 23.790.82 ± 0.040.90 ± 0.01
ShuffleNet86.13 ± 10.1683.54 ± 19.8989.05 ± 5.770.86 ± 0.120.93 ± 0.06
ResNet-1890.16 ± 2.3689.45 ± 7.3190.95 ± 9.290.91 ± 0.020.96 ± 0.05
ResNet-5092.62 ± 4.1991.14 ± 3.3594.29 ± 5.150.93 ± 0.040.98 ± 0.01
ResNet-10189.71 ± 10.0582.28 ± 20.0998.10 ± 2.180.89 ± 0.120.97 ± 0.03
Xception85.68 ± 6.7690.72 ± 4.7980.00 ± 19.640.87 ± 0.050.94 ± 0.04
Inception-v391.28 ± 8.2590.30 ± 5.1292.38 ± 11.980.92 ± 0.080.97 ± 0.02
Inception-ResNet-v286.35 ± 5.7188.19 ± 6.3784.29 ± 14.500.87 ± 0.050.95 ± 0.05
VGG-1678.52 ± 10.0274.68 ± 30.1482.86 ± 15.910.76 ± 0.170.91 ± 0.04
VGG-1983.22 ± 5.8590.72 ± 3.1974.76 ± 12.960.85 ± 0.040.90 ± 0.05
DenseNet-20191.72 ± 6.5288.61 ± 8.8695.24 ± 4.360.92 ± 0.070.97 ± 0.03
MobileNet-v287.25 ± 10.4695.78 ± 2.6477.62 ± 21.630.89 ± 0.080.95 ± 0.04
NasNet-Mobile83.45 ± 7.3684.81 ± 2.1981.90 ± 17.460.85 ± 0.050.94 ± 0.04
NasNet-Large85.23 ± 8.2579.32 ± 16.2891.90 ± 5.770.84 ± 0.100.93 ± 0.05
Table 3

Classification results without data augmentation.

CNN modelAccuracy (%)Sensitivity (%)Specificity (%)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$F_1$$\end{document}F1 scoreAUC
AlexNet86.85 ± 13.6680.25 ± 22.4994.29 ± 4.840.85 ± 0.160.94 ± 0.04
GoogLeNet93.83 ± 6.9796.71 ± 4.0690.57 ± 10.530.94 ± 0.060.96 ± 0.04
SqueezeNet87.52 ± 6.4586.84 ± 10.1188.29 ± 12.010.88 ± 0.060.94 ± 0.06
ShuffleNet95.97 ± 5.0995.44 ± 7.4796.57 ± 2.960.96 ± 0.050.97 ± 0.03
ResNet-1895.44 ± 8.0298.99 ± 1.6591.43 ± 15.250.96 ± 0.070.98 ± 0.03
ResNet-5093.62 ± 6.1795.57 ± 6.2791.43 ± 6.060.94 ± 0.060.98 ± 0.02
ResNet-10193.29 ± 5.6996.20 ± 1.7990.00 ± 10.100.94 ± 0.050.98 ± 0.02
Xception91.11 ± 10.1489.56 ± 12.5592.86 ± 7.800.91 ± 0.100.96 ± 0.03
Inception-v393.62 ± 5.2296.20 ± 0.0090.71 ± 11.110.94 ± 0.070.97 ± 0.04
Inception-ResNet-v288.59 ± 7.5989.24 ± 2.6987.86 ± 13.130.89 ± 0.070.96 ± 0.05
VGG-1689.26 ± 8.8092.83 ± 6.2485.24 ± 14.450.90 ± 0.080.96 ± 0.03
VGG-1990.16 ± 7.7287.34 ± 10.3693.33 ± 5.770.90 ± 0.080.97 ± 0.03
DenseNet-20196.20 ± 4.9595.78 ± 5.2796.67 ± 4.590.96 ± 0.050.98 ± 0.03
MobileNet-v295.97 ± 7.1896.71 ± 6.0495.14 ± 8.550.96 ± 0.070.97 ± 0.05
NasNet-Mobile89.26 ± 8.1491.56 ± 5.1286.67 ± 13.270.90 ± 0.070.95 ± 0.06
NasNet-Large88.59 ± 7.5990.51 ± 0.9086.43 ± 17.170.90 ± 0.060.96 ± 0.03
Classification results with data augmentation. Classification results without data augmentation. For the case of training the networks without data augmentation, DenseNet-201, MobileNet-v2, ShuffleNet, and ResNet-18 provide the average accuracy > 95%, while DenseNet-201 achieves the highest average accuracy (96.20%). GoogLeNet, ShuffleNet, ResNet-18, ResNet-50, ResNet-101, Inception-v3, DenseNet-201, and MobileNet-v2 result in the average sensitivity > 95%, while ResNet-18 has the highest average sensitivity (98.99%). ShuffleNet, DenseNet-201, and MobileNet-v2 provide the average specificity > 95%, while DenseNet-201 gives the highest average specificity (96.67%). ShuffleNet, ResNet-18, DenseNet-201, and MobileNet-v2 result in the highest average score = 0.96. The top three CNNs that achieve sensitivity > 95%, specificity > 95%, and score > 0.95 are ShuffleNet, DenseNet-201, and MobileNet-v2. DenseNet-201 is the best model for the classification of COVID-19 CT data. The networks that have the highest average AUC (0.98) are the ResNet family and DenseNet-201. The results obtained from CNNs without data augmentation are all better than those with data augmentation in terms of accuracy (compare results between Tables 2 and 3). For the case of training the networks with data augmentation, models that have average accuracy > 90% are ResNet-18 (90.16%), ResNet-50 (92.62%), Inception-v3 b(91.28%), and DenseNet-201 (91.72%); models that have average sensitivity > 90% are SqueezeNet (91.56%), ResNet-50 (91.14%), Xception (90.72%), Inception-v3 (90.30%), VGG-19 (90.72%), and MobileNet-v2 (95.78%); models that have average specificity > 90% are ResNet-18 (90.95%), ResNet-50 (94.29%), ResNet-101 (98.10%), Inception-v3 (92.38%), DenseNet-201 (95.24%), and NasNet-Large (91.90%); and models that have average score > 0.9 are ResNet-18 (0.91), ResNet-50 (0.93), Inception-v3 (0.92), and DenseNet-201 (0.92). Two CNNs using data augmentation that have accuracy > 90%, sensitivity > 90%, specificity > 90%, and score > 0.9 are ResNet-50 and Inception-v3. The networks without data augmenttaion have higher or equal values for the average AUC than or to those with data augmentation. Plots of accuracy vs. relative training time (ratio of training time of a network to the training time of the SqueezeNet) of 16 pretrained CNNs using COVID-19 CT database, where the circle size indicates the magnitude of memory in MB. In summary, without data augmentation, the best classifier is DenseNet-201, which has the best accuracy, best balance between sensitivity and specificity, top score, and top AUC. Figure 2 shows the plot of accuracy versus relative training time obtained from the 16 pretrained CNNs without data augmentation.
Figure 2

Plots of accuracy vs. relative training time (ratio of training time of a network to the training time of the SqueezeNet) of 16 pretrained CNNs using COVID-19 CT database, where the circle size indicates the magnitude of memory in MB.

Discussion

The benchmark results using the same database reported in[12], with a fixed split data of about 80% for training and 20% for testing, have accuracy = 84.7%, sensitivity = 76.2%, and score = 0.85, using a fine-tuned pretrained DenseNet with data augmentation. The results obtained from the 16 CNNs without data augmentation are better than these benchmark results. The study published in[13] applied 10 pretrained CNN using a different COVID-19 database with the same ratio of training and testing data, which is not publicly available, reported among all the 10 networks, ResNet-101 was the best classification model. ResNet-101 achieved accuracy = 99.51%, sensitivity = 100%, and specificity = 99.02%. Although using a different database, the results obtained in this study are comparable. However, the input data processing and training reported in[13] requires much effort by requiring the extraction of regions of interest by a radiologist, which is subjective, time-consuming, and likely hinders the real-time application of the pretrained networks. The work reported in[14] requires the pre-processing of 3D CT scans by extracting the regions of interest using a U-Net for image segmentation. The pre-processed images were then passed to the COVNet for the prediction. The sensitivity and specificity obtained from COVNet were 87% and 92%, respectively, using a dataset that is not publicly available. Another work on the classification of COVID-19 CT images collected from 259 patients reported in[15] modified the pretrained Inception that achieved accuracy = 79.3%, sensitivity = 67%, and specificity = 83%, and another test achieving accuracy = 85.2%. Similarly, the input images are extracted regions of interest such as small patchy shadows and interstitial changes, multiple ground glass and infiltrates in both lungs. The study reported in[16] used the concatenation of two pretrained ResNet-based networks and the Bayesian function for screening COVID-19 patients using CT imaging. The data pre-processing of classification procedure requires 3D segmentation, extraction of regions of interest (such as ground-glass appearance, striking peripheral distribution along with the pleura, and independent focus of infections), and data augmentation. The overall accuracy obtained was 86.7%. The classification results obtained in this study are preferable to those reported in[14-16] in terms of accuracy and implementation of input data. Although the use of regions of interest or cropped images is widely adopted for deep learning, including other classification problems[19-22], this study finds that the direct input of CT images, which are then resized to fit the input size of the pretrained CNN, and transfer learning without data augmentation can achieve very high and better classification performance than those using data augmentation. Such findings are useful for the rapid deployment of AI tools to meet the urgent demand for curbing the pandemic, because it can relieve the task of manual detection of regions of interest carried out by experienced radiologists, employment of image segmentation methods, and more data collection. A training process of DenseNet-201. Using the described network-training configuration with only 6 epochs, the CNNs could provide a very high performance of classification. Figures 3 and 4 show one of the training processes of DenseNet-201 (best network) and some features obtained from the deep learning of the best network, respectively.
Figure 3

A training process of DenseNet-201.

Figure 4

Features learned by DenseNet-201: 36 features in layer ’conv1|conv’(convolution) (top left), layer ‘conv4_block7_1_conv’ (convolution) (top right), layer ‘conv5_block9_1_conv’ (convolution) (bottom left), and 2 features in layer ‘new_fc’ (fully connected) (bottom right).

Features learned by DenseNet-201: 36 features in layer ’conv1|conv’(convolution) (top left), layer ‘conv4_block7_1_conv’ (convolution) (top right), layer ‘conv5_block9_1_conv’ (convolution) (bottom left), and 2 features in layer ‘new_fc’ (fully connected) (bottom right). As the numbers of COVID-19 and non-COVID-19 CT images used in this study are 349 and 397, respectively, the binary classification in this study was not much disadvantaged from the class imbalance problem, where the class distributions are highly imbalanced. Due to imbalanced data, classifiers tend to result in low predictive accuracy for the minority class. Medical datasets are often not balanced in the class labels because of limited samples collected from patients and cost for acquiring annotated data. There are many techniques proposed for addressing class imbalance , which can be applied to medical imaging, such as the “deep domain adaptation”[23] for handling the shortage of large amounts of labeled data, weighted loss method by updating the loss function to result in the same loss for all classes, downsampling by removing images from the majority class, and oversampling by adding more images to minority classes using artificial data augmentation[24,25]. Open challenges in imbalance data and exploration for solutions can be found in[26].

Conclusions

AI-based medical diagnosis systems based on deep learning of medical imaging are increasingly recognized to be clinically useful. However, development of suitable deep-learning networks and effective training strategy for clinical applications is a topic of research that needs to be explored[27]. Through a comprehensive investigation of 16 pretrained CNNs using certain parameter specification and training strategy for the networks, this study discovers the very high performance of several of these networks for COVID-19 diagnosis using CT images. The network configuration of the pretrained models can be implemented for classification of other image modality, such as X-ray, for the detection of COVID-19. Most AI studies on chest CT used for differentiating COVID-19 pneumonia from other causes of pneumonia consider both three-class classification problems (COVID-19 pneumonia, non-COVID-19 pneumonia, and healthy) and two-class classification (COVID-19 pneumonia and healthy)[2]. Due to the limit of publicly available data, this study concerns with the two-class classification. However, extension of the use of pretrained CNNs to the three-class classification of COVID-19 imaging data is straightforward. The findings reported from this study bring benefits to the development of fast and efficient diagnostic tools using imaging data and contribute to further leading into the development of more accurate point-of-care diagnostic and detection tools for containing the coronavirus pandemic.
  12 in total

1.  Time-Independent Prediction of Burn Depth Using Deep Convolutional Neural Networks.

Authors:  Marco Domenico Cirillo; Robin Mirdell; Folke Sjöberg; Tuan D Pham
Journal:  J Burn Care Res       Date:  2019-10-16       Impact factor: 1.845

2.  Transfer learning using a multi-scale and multi-network ensemble for skin lesion classification.

Authors:  Amirreza Mahbod; Gerald Schaefer; Chunliang Wang; Georg Dorffner; Rupert Ecker; Isabella Ellinger
Journal:  Comput Methods Programs Biomed       Date:  2020-03-21       Impact factor: 5.428

3.  A transfer learning method with deep residual network for pediatric pneumonia diagnosis.

Authors:  Gaobo Liang; Lixin Zheng
Journal:  Comput Methods Programs Biomed       Date:  2019-06-26       Impact factor: 5.428

4.  Coronavirus disease 2019 (COVID-19) imaging reporting and data system (COVID-RADS) and common lexicon: a proposal based on the imaging data of 37 studies.

Authors:  Sana Salehi; Aidin Abedi; Sudheer Balakrishnan; Ali Gholamrezanezhad
Journal:  Eur Radiol       Date:  2020-04-28       Impact factor: 5.315

5.  Application of deep learning technique to manage COVID-19 in routine clinical practice using CT images: Results of 10 convolutional neural networks.

Authors:  Ali Abbasian Ardakani; Alireza Rajabzadeh Kanafi; U Rajendra Acharya; Nazanin Khadem; Afshin Mohammadi
Journal:  Comput Biol Med       Date:  2020-04-30       Impact factor: 4.589

6.  CT imaging changes of corona virus disease 2019(COVID-19): a multi-center study in Southwest China.

Authors:  Xiaoming Li; Wenbing Zeng; Xiang Li; Haonan Chen; Linping Shi; Xinghui Li; Hongnian Xiang; Yang Cao; Hui Chen; Chen Liu; Jian Wang
Journal:  J Transl Med       Date:  2020-04-06       Impact factor: 5.531

7.  CT in coronavirus disease 2019 (COVID-19): a systematic review of chest CT findings in 4410 adult patients.

Authors:  Vineeta Ojha; Avinash Mani; Niraj Nirmal Pandey; Sanjiv Sharma; Sanjeev Kumar
Journal:  Eur Radiol       Date:  2020-05-30       Impact factor: 5.315

8.  COVID-19 and artificial intelligence: protecting health-care workers and curbing the spread.

Authors:  Becky McCall
Journal:  Lancet Digit Health       Date:  2020-02-20

9.  Using Artificial Intelligence to Detect COVID-19 and Community-acquired Pneumonia Based on Pulmonary CT: Evaluation of the Diagnostic Accuracy.

Authors:  Lin Li; Lixin Qin; Zeguo Xu; Youbing Yin; Xin Wang; Bin Kong; Junjie Bai; Yi Lu; Zhenghan Fang; Qi Song; Kunlin Cao; Daliang Liu; Guisheng Wang; Qizhong Xu; Xisheng Fang; Shiqin Zhang; Juan Xia; Jun Xia
Journal:  Radiology       Date:  2020-03-19       Impact factor: 11.105

Review 10.  The role of imaging in 2019 novel coronavirus pneumonia (COVID-19).

Authors:  Wenjing Yang; Arlene Sirajuddin; Xiaochun Zhang; Guanshu Liu; Zhongzhao Teng; Shihua Zhao; Minjie Lu
Journal:  Eur Radiol       Date:  2020-04-15       Impact factor: 5.315

View more
  26 in total

1.  Non-iterative learning machine for identifying CoViD19 using chest X-ray images.

Authors:  Sahil Dalal; Virendra P Vishwakarma; Varsha Sisaudia; Parul Narwal
Journal:  Sci Rep       Date:  2022-07-13       Impact factor: 4.996

2.  Helping Roles of Artificial Intelligence (AI) in the Screening and Evaluation of COVID-19 Based on the CT Images.

Authors:  Hui Xie; Qing Li; Ping-Feng Hu; Sen-Hua Zhu; Jian-Fang Zhang; Hong-Da Zhou; Hai-Bo Zhou
Journal:  J Inflamm Res       Date:  2021-03-26

Review 3.  Transfer learning for medical image classification: a literature review.

Authors:  Mate E Maros; Thomas Ganslandt; Hee E Kim; Alejandro Cosa-Linan; Nandhini Santhanam; Mahboubeh Jannesari
Journal:  BMC Med Imaging       Date:  2022-04-13       Impact factor: 1.930

4.  Mutual stain conversion between Giemsa and Papanicolaou in cytological images using cycle generative adversarial network.

Authors:  Atsushi Teramoto; Ayumi Yamada; Tetsuya Tsukamoto; Yuka Kiriyama; Eiko Sakurai; Kazuya Shiogama; Ayano Michiba; Kazuyoshi Imaizumi; Kuniaki Saito; Hiroshi Fujita
Journal:  Heliyon       Date:  2021-02-24

5.  The prediction of the lifetime of the new coronavirus in the USA using mathematical models.

Authors:  K Selvakumar; S Lokesh
Journal:  Soft comput       Date:  2021-03-10       Impact factor: 3.732

6.  Machine learning models for predicting critical illness risk in hospitalized patients with COVID-19 pneumonia.

Authors:  Qin Liu; Baoguo Pang; Haijun Li; Bin Zhang; Yumei Liu; Lihua Lai; Wenjun Le; Jianyu Li; Tingting Xia; Xiaoxian Zhang; Changxing Ou; Jianjuan Ma; Shenghao Li; Xiumei Guo; Shuixing Zhang; Qingling Zhang; Min Jiang; Qingsi Zeng
Journal:  J Thorac Dis       Date:  2021-02       Impact factor: 2.895

7.  Risk of in-hospital death associated with Covid-19 lung consolidations on chest computed tomography - A novel translational approach using a radiation oncology contour software.

Authors:  Lucas G Sapienza; Karim Nasra; Vinícius F Calsavara; Tania B Little; Vrinda Narayana; Eyad Abu-Isa
Journal:  Eur J Radiol Open       Date:  2021-01-06

8.  Efficient deep neural networks for classification of COVID-19 based on CT images: Virtualization via software defined radio.

Authors:  Saman Fouladi; M J Ebadi; Ali A Safaei; Mohd Yazid Bajuri; Ali Ahmadian
Journal:  Comput Commun       Date:  2021-06-16       Impact factor: 3.167

Review 9.  AI for COVID-19 Detection from Radiographs: Incisive Analysis of State of the Art Techniques, Key Challenges and Future Directions.

Authors:  R Karthik; R Menaka; M Hariharan; G S Kathiresan
Journal:  Ing Rech Biomed       Date:  2021-07-26

Review 10.  Viral outbreaks detection and surveillance using wastewater-based epidemiology, viral air sampling, and machine learning techniques: A comprehensive review and outlook.

Authors:  Omar M Abdeldayem; Areeg M Dabbish; Mahmoud M Habashy; Mohamed K Mostafa; Mohamed Elhefnawy; Lobna Amin; Eslam G Al-Sakkari; Ahmed Ragab; Eldon R Rene
Journal:  Sci Total Environ       Date:  2021-08-21       Impact factor: 7.963

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.