| Literature DB >> 33169050 |
Yu-Dong Zhang1,2, Suresh Chandra Satapathy3, Shuaiqi Liu4, Guang-Run Li5.
Abstract
Till August 17, 2020, COVID-19 has caused 21.59 million confirmed cases in more than 227 countries and territories, and 26 naval ships. Chest CT is an effective way to detect COVID-19. This study proposed a novel deep learning model that can diagnose COVID-19 on chest CT more accurately and swiftly. Based on traditional deep convolutional neural network (DCNN) model, we proposed three improvements: (i) We introduced stochastic pooling to replace average pooling and max pooling; (ii) We combined conv layer with batch normalization layer and obtained the conv block (CB); (iii) We combined dropout layer with fully connected layer and obtained the fully connected block (FCB). Our algorithm achieved a sensitivity of 93.28% ± 1.50%, a specificity of 94.00% ± 1.56%, and an accuracy of 93.64% ± 1.42%, in identifying COVID-19 from normal subjects. We proved using stochastic pooling yields better performance than average pooling and max pooling. We compared different structure configurations and proved our 3CB + 2FCB yields the best performance. The proposed model is effective in detecting COVID-19 based on chest CT images. © Springer-Verlag GmbH Germany, part of Springer Nature 2020.Entities:
Keywords: Batch normalization; COVID-19; Convolution block; Deep convolutional neural network; Dropout; Fully connected block; Stochastic pooling
Year: 2020 PMID: 33169050 PMCID: PMC7609373 DOI: 10.1007/s00138-020-01128-8
Source DB: PubMed Journal: Mach Vis Appl ISSN: 0932-8092 Impact factor: 2.012
Demographics of COVID-19 and HC
| No. Subjects | No. Images | Age Range | |
|---|---|---|---|
| COVID-19 | 142 | 320 | 22–91 |
| HC | 142 | 320 | 21–76 |
Abbreviation list
| Meanings | Abbreviations |
|---|---|
| CCT | Chest computed tomography |
| BCR | Byte compression ratio |
| SLS | Slice level selection |
| NLAF | Nonlinear activation function |
| AM | Activation map |
| (A)(M)(S)P | (average) (max) (stochastic) pooling |
| NLDS | nonlinear downsampling |
| DW | Down-weight |
| DO(L)(N) | Dropout (layer) (neuron) |
| CRLW | Compression ratio of learnable weights |
| PL | Pooling layer |
| SC | Structure configuration |
| CB | Convolution block |
| FCB | Fully connected block |
Fig. 1Diagram of preprocessing (color figure online)
Image size and storage per image at each preprocessing step
| Preprocessing step | Image Size (per image) | Byte(s) (per image) |
|---|---|---|
| Original | 12,582,912 | |
| Grayscaled | 4,194,304 | |
| Histogram stretched | 4,194,304 | |
| Cropped | 2,096,704 | |
| Downsampled | 65,536 |
Fig. 2Two samples of our preprocessed dataset
Fig. 3Pipeline of a toy example of DCNN with 2 CLs, 2PLs, and 2 FCLs
Fig. 4Pipeline of conv layer
Fig. 5Toy examples of different pooling technologies
Fig. 6DONs at training and inference stages ( = weights, =retention probability)
Fig. 7A toy example of a DCNN with four FCLs
Fig. 8Structure of proposed 5L-DCNN-SP-C
Details of each layer in proposed 5L-DCNN-SP-C
| Layer/Block | Hyperparameter | AM |
|---|---|---|
| Input | n/a | 128 × 128 × 1 |
| CB-1-SP | 32 3 × 3 /2 | 64 × 64 × 32 |
| CB-2-SP | 64 3 × 3 /2 | 32 × 32 × 64 |
| CB-3-SP | 128 3 × 3 /2 | 16 × 16 × 128 |
| Flatten | 1 × 32,768 | |
| FCB-4 | W(50 × 32,768); B(50 × 1); | 1 × 50 |
| FCB-5 | W(2 × 50); B(2 × 1); | 1 × 2 |
n/a not available, AM activation map
Ten runs of AP, MP, and SP
| AP | ||||||
|---|---|---|---|---|---|---|
| 1 | 93.13 | 90.00 | 90.30 | 91.56 | 91.69 | 83.17 |
| 2 | 90.94 | 90.00 | 90.09 | 90.47 | 90.51 | 80.94 |
| 3 | 92.50 | 90.94 | 91.08 | 91.72 | 91.78 | 83.45 |
| 4 | 90.94 | 92.19 | 92.09 | 91.56 | 91.51 | 83.13 |
| 5 | 91.56 | 90.63 | 90.71 | 91.09 | 91.14 | 82.19 |
| 6 | 91.88 | 92.50 | 92.45 | 92.19 | 92.16 | 84.38 |
| 7 | 90.94 | 90.63 | 90.65 | 90.78 | 90.80 | 81.56 |
| 8 | 92.19 | 89.38 | 89.67 | 90.78 | 90.91 | 81.59 |
| 9 | 88.75 | 90.31 | 90.16 | 89.53 | 89.45 | 79.07 |
| 10 | 89.38 | 88.13 | 88.27 | 88.75 | 88.82 | 77.51 |
| Mean ± SD | 91.22 ± 1.35 | 90.47 ± 1.27 | 90.55 ± 1.19 | 90.84 ± 1.05 | 90.88 ± 1.06 | 81.70 ± 2.10 |
Fig. 9Error bar of different pooling methods
SC setting
| SC | ||
|---|---|---|
| I | 2 | 1 |
| II | 2 | 2 |
| III | 2 | 3 |
| IV | 3 | 1 |
| V (Ours) | 3 | 2 |
| VI | 3 | 3 |
SC structure configuration, number of CBs, number of FCBs
Performances of all six SCs (bold means the best)
| SC | ||||||
|---|---|---|---|---|---|---|
| I | 92.28 ± 0.69 | 91.00 ± 1.53 | 91.13 ± 1.39 | 91.64 ± 0.86 | 91.70 ± 0.82 | 83.30 ± 1.72 |
| II | 93.13 ± 1.32 | 92.59 ± 1.36 | 92.65 ± 1.26 | 92.86 ± 1.00 | 92.88 ± 1.00 | 85.73 ± 1.99 |
| III | 93.28 ± 0.61 | 92.97 ± 1.33 | 93.01 ± 1.23 | 93.13 ± 0.71 | 93.14 ± 0.68 | 86.26 ± 1.43 |
| IV | 92.69 ± 1.11 | 92.53 ± 1.88 | 92.57 ± 1.71 | 92.61 ± 1.08 | 92.62 ± 1.04 | 85.24 ± 2.14 |
| V (Ours) | 93.28 ± 1.50 | |||||
| VI | 93.03 ± 1.09 | 93.07 ± 0.98 | 93.23 ± 0.82 | 93.24 ± 0.85 | 86.49 ± 1.65 |
Comparison with SOTA approaches (Unit: %)
| Approach | ||||||
|---|---|---|---|---|---|---|
| RBFNN [ | 67.08 | 74.48 | 72.52 | 70.78 | 69.64 | 41.74 |
| K-ELM [ | 57.29 | 61.46 | 59.83 | 59.38 | 58.46 | 18.81 |
| ELM-BA [ | 57.08 ± 3.86 | 72.40 ± 3.03 | 67.48 ± 1.65 | 64.74 ± 1.26 | 61.75 ± 2.24 | 29.90 ± 2.45 |
| 6L-CNN-F [ | 81.04 ± 2.90 | 79.27 ± 2.21 | 79.70 ± 1.27 | 80.16 ± 0.85 | 80.31 ± 1.13 | 60.42 ± 1.73 |
| GoogLeNet [ | 76.88 ± 3.92 | 83.96 ± 2.29 | 82.84 ± 1.58 | 80.42 ± 1.40 | 79.65 ± 1.92 | 61.10 ± 2.62 |
| ResNet-18 [ | 78.96 ± 2.90 | 89.48 ± 1.64 | 88.30 ± 1.50 | 84.22 ± 1.23 | 83.31 ± 1.53 | 68.89 ± 2.33 |
5L-DCNN-SP-C (Ours) | 93.28 ± 1.50 | 94.00 ± 1.56 | 93.96 ± 1.54 | 93.64 ± 1.42 | 93.62 ± 1.42 | 87.29 ± 2.83 |
Fig. 10Comparison to state-of-the-art approaches