| Literature DB >> 33037291 |
Abstract
The use of imaging data has been reported to be useful for rapid diagnosis of COVID-19. Although computed tomography (CT) scans show a variety of signs caused by the viral infection, given a large amount of images, these visual features are difficult and can take a long time to be recognized by radiologists. Artificial intelligence methods for automated classification of COVID-19 on CT scans have been found to be very promising. However, current investigation of pretrained convolutional neural networks (CNNs) for COVID-19 diagnosis using CT data is limited. This study presents an investigation on 16 pretrained CNNs for classification of COVID-19 using a large public database of CT scans collected from COVID-19 patients and non-COVID-19 subjects. The results show that, using only 6 epochs for training, the CNNs achieved very high performance on the classification task. Among the 16 CNNs, DenseNet-201, which is the deepest net, is the best in terms of accuracy, balance between sensitivity and specificity, [Formula: see text] score, and area under curve. Furthermore, the implementation of transfer learning with the direct input of whole image slices and without the use of data augmentation provided better classification rates than the use of data augmentation. Such a finding alleviates the task of data augmentation and manual extraction of regions of interest on CT images, which are adopted by current implementation of deep-learning models for COVID-19 classification.Entities:
Mesh:
Year: 2020 PMID: 33037291 PMCID: PMC7547710 DOI: 10.1038/s41598-020-74164-z
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1CT images: Rows 1 and 2: COVID-19, Rows 3 and 4: non-COVID-19 (aspect ratios of some images were rescaled to fit the figure frame).
Properties of 16 pre-trained CNNs.
| CNN | Depth | Size (MB) | Parameters (millions) | Input image size |
|---|---|---|---|---|
| AlexNet | 8 | 227 | 61.0 | 227 |
| GoogLeNet | 22 | 27 | 7.0 | 224 |
| SqueezeNet | 18 | 4.6 | 1.24 | 227 |
| ShuffleNet | 50 | 6.3 | 1.4 | 224 |
| ResNet-18 | 18 | 44 | 11.7 | 224 |
| ResNet-50 | 50 | 96 | 25.6 | 224 |
| ResNet-101 | 101 | 167 | 44.6 | 224 |
| Xception | 71 | 85 | 22.9 | 299 |
| Inception-v3 | 48 | 89 | 23.9 | 299 |
| Inception-ResNet-v2 | 164 | 209 | 55.9 | 299 |
| VGG-16 | 16 | 515 | 138 | 224 |
| VGG-19 | 19 | 535 | 144 | 224 |
| DenseNet-201 | 201 | 77 | 20.0 | 224 |
| MobileNet-v2 | 53 | 13 | 3.5 | 224 |
| NasNet-Mobile | * | 20 | 5.3 | 224 |
| NasNet-Large | * | 360 | 88.9 | 331 |
*indicates NASNet-Mobile and NasNetLarge networks do not consist of a linear sequence of modules.
Classification results with data augmentation.
| CNN model | Accuracy (%) | Sensitivity (%) | Specificity (%) | AUC | |
|---|---|---|---|---|---|
| AlexNet | 74.50 ± 4.40 | 70.46 ± 6.37 | 79.05 ± 8.61 | 0.75 ± 0.04 | 0.83 ± 0.04 |
| GoogLeNet | 78.97 ± 3.70 | 75.95 ± 13.69 | 82.38 ± 10.53 | 0.79 ± 0.06 | 0.91 ± 0.04 |
| SqueezeNet | 78.52 ± 7.56 | 91.56 ± 7.63 | 63.81 ± 23.79 | 0.82 ± 0.04 | 0.90 ± 0.01 |
| ShuffleNet | 86.13 ± 10.16 | 83.54 ± 19.89 | 89.05 ± 5.77 | 0.86 ± 0.12 | 0.93 ± 0.06 |
| ResNet-18 | 90.16 ± 2.36 | 89.45 ± 7.31 | 90.95 ± 9.29 | 0.91 ± 0.02 | 0.96 ± 0.05 |
| ResNet-50 | 92.62 ± 4.19 | 91.14 ± 3.35 | 94.29 ± 5.15 | 0.93 ± 0.04 | 0.98 ± 0.01 |
| ResNet-101 | 89.71 ± 10.05 | 82.28 ± 20.09 | 98.10 ± 2.18 | 0.89 ± 0.12 | 0.97 ± 0.03 |
| Xception | 85.68 ± 6.76 | 90.72 ± 4.79 | 80.00 ± 19.64 | 0.87 ± 0.05 | 0.94 ± 0.04 |
| Inception-v3 | 91.28 ± 8.25 | 90.30 ± 5.12 | 92.38 ± 11.98 | 0.92 ± 0.08 | 0.97 ± 0.02 |
| Inception-ResNet-v2 | 86.35 ± 5.71 | 88.19 ± 6.37 | 84.29 ± 14.50 | 0.87 ± 0.05 | 0.95 ± 0.05 |
| VGG-16 | 78.52 ± 10.02 | 74.68 ± 30.14 | 82.86 ± 15.91 | 0.76 ± 0.17 | 0.91 ± 0.04 |
| VGG-19 | 83.22 ± 5.85 | 90.72 ± 3.19 | 74.76 ± 12.96 | 0.85 ± 0.04 | 0.90 ± 0.05 |
| DenseNet-201 | 91.72 ± 6.52 | 88.61 ± 8.86 | 95.24 ± 4.36 | 0.92 ± 0.07 | 0.97 ± 0.03 |
| MobileNet-v2 | 87.25 ± 10.46 | 95.78 ± 2.64 | 77.62 ± 21.63 | 0.89 ± 0.08 | 0.95 ± 0.04 |
| NasNet-Mobile | 83.45 ± 7.36 | 84.81 ± 2.19 | 81.90 ± 17.46 | 0.85 ± 0.05 | 0.94 ± 0.04 |
| NasNet-Large | 85.23 ± 8.25 | 79.32 ± 16.28 | 91.90 ± 5.77 | 0.84 ± 0.10 | 0.93 ± 0.05 |
Classification results without data augmentation.
| CNN model | Accuracy (%) | Sensitivity (%) | Specificity (%) | AUC | |
|---|---|---|---|---|---|
| AlexNet | 86.85 ± 13.66 | 80.25 ± 22.49 | 94.29 ± 4.84 | 0.85 ± 0.16 | 0.94 ± 0.04 |
| GoogLeNet | 93.83 ± 6.97 | 96.71 ± 4.06 | 90.57 ± 10.53 | 0.94 ± 0.06 | 0.96 ± 0.04 |
| SqueezeNet | 87.52 ± 6.45 | 86.84 ± 10.11 | 88.29 ± 12.01 | 0.88 ± 0.06 | 0.94 ± 0.06 |
| ShuffleNet | 95.97 ± 5.09 | 95.44 ± 7.47 | 96.57 ± 2.96 | 0.96 ± 0.05 | 0.97 ± 0.03 |
| ResNet-18 | 95.44 ± 8.02 | 98.99 ± 1.65 | 91.43 ± 15.25 | 0.96 ± 0.07 | 0.98 ± 0.03 |
| ResNet-50 | 93.62 ± 6.17 | 95.57 ± 6.27 | 91.43 ± 6.06 | 0.94 ± 0.06 | 0.98 ± 0.02 |
| ResNet-101 | 93.29 ± 5.69 | 96.20 ± 1.79 | 90.00 ± 10.10 | 0.94 ± 0.05 | 0.98 ± 0.02 |
| Xception | 91.11 ± 10.14 | 89.56 ± 12.55 | 92.86 ± 7.80 | 0.91 ± 0.10 | 0.96 ± 0.03 |
| Inception-v3 | 93.62 ± 5.22 | 96.20 ± 0.00 | 90.71 ± 11.11 | 0.94 ± 0.07 | 0.97 ± 0.04 |
| Inception-ResNet-v2 | 88.59 ± 7.59 | 89.24 ± 2.69 | 87.86 ± 13.13 | 0.89 ± 0.07 | 0.96 ± 0.05 |
| VGG-16 | 89.26 ± 8.80 | 92.83 ± 6.24 | 85.24 ± 14.45 | 0.90 ± 0.08 | 0.96 ± 0.03 |
| VGG-19 | 90.16 ± 7.72 | 87.34 ± 10.36 | 93.33 ± 5.77 | 0.90 ± 0.08 | 0.97 ± 0.03 |
| DenseNet-201 | 96.20 ± 4.95 | 95.78 ± 5.27 | 96.67 ± 4.59 | 0.96 ± 0.05 | 0.98 ± 0.03 |
| MobileNet-v2 | 95.97 ± 7.18 | 96.71 ± 6.04 | 95.14 ± 8.55 | 0.96 ± 0.07 | 0.97 ± 0.05 |
| NasNet-Mobile | 89.26 ± 8.14 | 91.56 ± 5.12 | 86.67 ± 13.27 | 0.90 ± 0.07 | 0.95 ± 0.06 |
| NasNet-Large | 88.59 ± 7.59 | 90.51 ± 0.90 | 86.43 ± 17.17 | 0.90 ± 0.06 | 0.96 ± 0.03 |
Figure 2Plots of accuracy vs. relative training time (ratio of training time of a network to the training time of the SqueezeNet) of 16 pretrained CNNs using COVID-19 CT database, where the circle size indicates the magnitude of memory in MB.
Figure 3A training process of DenseNet-201.
Figure 4Features learned by DenseNet-201: 36 features in layer ’conv1|conv’(convolution) (top left), layer ‘conv4_block7_1_conv’ (convolution) (top right), layer ‘conv5_block9_1_conv’ (convolution) (bottom left), and 2 features in layer ‘new_fc’ (fully connected) (bottom right).