| Literature DB >> 33488837 |
Yu-Dong Zhang1, Suresh Chandra Satapathy2, Xin Zhang3, Shui-Hua Wang4,5.
Abstract
COVID-19 is an ongoing pandemic disease. To make more accurate diagnosis on COVID-19 than existing approaches, this paper proposed a novel method combining DenseNet and optimization of transfer learning setting (OTLS) strategy. Preprocessing was used to enhance, crop, and resize the collected chest CT images. Data augmentation method was used to increase the size of training set. A composite learning factor (CLF) was employed which assigned different learning factor to three types of layers: frozen layers, middle layers, and new layers. Meanwhile, the OTLS strategy was proposed. Finally, precomputation method was utilized to reduce RAM storage and accelerate the algorithm. We observed that optimization setting "201-IV" can achieve the best performance by proposed OTLS strategy. The sensitivity, specificity, precision, and accuracy of our proposed method were 96.35 ± 1.07, 96.25 ± 1.16, 96.29 ± 1.11, and 96.30 ± 0.56, respectively. The proposed DenseNet-OTLS method achieved better performances than state-of-the-art approaches in diagnosing COVID-19. © Springer Science+Business Media, LLC, part of Springer Nature 2021.Entities:
Keywords: COVID-19; Composite learning factor; Data augmentation; DenseNet; Optimization; Precomputation; Transfer learning
Year: 2021 PMID: 33488837 PMCID: PMC7812362 DOI: 10.1007/s12559-020-09776-8
Source DB: PubMed Journal: Cognit Comput ISSN: 1866-9956 Impact factor: 5.418
Demographic statistics of subjects
| Subject Number | Image Number | |
|---|---|---|
| COVID-19 | 142 (95/47) | 320 |
| HC | 142 (88/54) | 320 |
Abbreviations and their full names
| Abbreviation | Full name |
|---|---|
| AP | Average pooling |
| BM | Base model |
| CLF | Composite learning factor |
| CLFS | Composite learning factor setting |
| CNN | Convolutional neural network |
| CT | Computed Tomography |
| DA | Data augmentation |
| DB | DenseBlock |
| FCL | Fully connected layer |
| FL | Frozen layer |
| GAP | Global average pooling |
| GGO | Ground-glass opacity |
| GLR | Global learning rate |
| HC | Healthy control |
| HS | Histogram stretching |
| HU | Hounsfield unit |
| ILSVRC | ImageNet large scale visual recognition challenge |
| LF | Learning factor |
| LR | Learning rate |
| LRN | Local response normalization |
| MCC | Matthews correlation coefficient |
| ML | Middle layer |
| MP | Maximum pooling |
| MV | Majority voting |
| NL | New layer |
| OTLS | Optimization of transfer learning setting |
| PCR | Polymerase chain reaction |
| PTM | Pre-trained model |
| RHO | Random hold-out |
| SD | Standard deviation |
| SGDM | Stochastic gradient descent with momentum |
| SLF | Simple learning factor |
| SPFP | Single-precision floating-point |
| TL | Transfer learning |
| TL | Transition layer |
Fig. 1Why we need to crop the chest CT image (the right and bottom show unrelated contents, outlined by red boxes)
Fig. 2Illustration of preprocessed COVID-19 dataset
RHO setting
| Set | COVID-19 | Healthy | Total |
|---|---|---|---|
| Training (50%) | 160 | 160 | 320 |
| Validation (20%) | 64 | 64 | 128 |
| Test (30%) | 96 | 96 | 192 |
| Total | 320 | 320 | 640 |
Fig. 3Data-augmented training samples (mirror results were not shown)
Fig. 4Idea of transfer learning
Fig. 5Comparison of plain CNN block, ResNet block, and DenseNet block
Fig. 6How DenseNet classify chest CT images (TL transition layer, DB DenseBlock, GAP global average pooling, FCL fully connected layer)
Modification of layers of DenseNet
| Layer | Original | Replaced |
|---|---|---|
| Third from last | FCL (1000) with pre-trained weights and biases | FCL (2) with random initialization |
| Last | Classification Layer 1000 classes: recreational vehicle, printer, coho, milk can, Irish wolfhound, parallel bars, tree frog, dhole, Gila monster, toucan, spider web, organ, walking stick, broccoli, loggerhead, bassoon, colobus, racket, schooner, and Kerry blue terrier, and 980 other classes | Classification Layer Two classes: (i) COVID-19; (ii) Healthy control |
Fig. 7Structure of DenseNet-121 (CP means the first block of conv layer and pooling layer, DB means the dense block, TL means the transition layer, GAP means global average pooling, FCL means fully connected layer, DL means dense layer)
Detailed information of each layer/block in DenseNet-121
| Index | Layers | DenseNet-121 | Output |
|---|---|---|---|
| 1 | Input | 224 × 224 × 3 | |
| 2 | Conv | 7 × 7/2 conv, | 112 × 112 × 64 |
| 3 | Pooling | 3 × 3/2 MP | 56 × 56 × 64 |
| 4 | DB1 | 56 × 56 × 256 | |
| 5 | TL1 | 1 × 1 conv, 2 × 2/2 AP | 28 × 28 × 128 |
| 6 | DB2 | 28 × 28 × 512 | |
| 7 | TL2 | 1 × 1 conv, 2 × 2/2 AP | 14 × 14 × 256 |
| 8 | DB3 | 14 × 14 × 1024 | |
| 9 | TL3 | 1 × 1 conv, 2 × 2/2 AP | 7 × 7 × 512 |
| 10 | DB4 | 7 × 7 × 1024 | |
| 11 | Pooling | 7 × 7/7 GAP | 1 × 1 × 1024 |
| 12 | FCL | 1000 days | 1 × 1 × 1000 |
DB dense block, TL transition layer, FCL fully connected layer
Composite-learning factor setting
| CLFS | Frozen layers | Middle layers | New layers |
|---|---|---|---|
| I | CP, DB1, TL1, DB2 | TL2, DB3, TL3, DB4 | FCL |
| II | CP, DB1, TL1, DB2, TL2 | DB3, TL3, DB4 | FCL |
| III | CP, DB1, TL1, DB2, TL2, DB3 | TL3, DB4 | FCL |
| IV | CP, DB1, TL1, DB2, TL2, DB3, TL3 | DB4 | FCL |
Fig. 8CLF Setting
Fig. 9Storage comparison of four transfer learning settings in DenseNet-201
Pseudocode of our proposed DenseNet-OTLS method
Validation performance based on the best configuration (unit: %)
| Run | Sen | Spc | Prc | Acc | F1 | MCC |
|---|---|---|---|---|---|---|
| R1 | 98.44 | 96.88 | 96.92 | 97.66 | 97.67 | 95.34 |
| R2 | 93.75 | 98.44 | 98.39 | 96.09 | 96.01 | 92.30 |
| R3 | 98.44 | 93.75 | 94.03 | 96.09 | 96.18 | 92.30 |
| R4 | 96.88 | 96.88 | 96.88 | 96.88 | 96.88 | 93.75 |
| R5 | 98.44 | 95.31 | 95.45 | 96.88 | 96.92 | 93.80 |
| R6 | 98.44 | 98.44 | 98.44 | 98.44 | 98.44 | 96.88 |
| R7 | 95.31 | 98.44 | 98.44 | 96.88 | 96.82 | 93.84 |
| R8 | 93.75 | 96.88 | 96.77 | 95.31 | 95.21 | 90.71 |
| R9 | 96.88 | 96.88 | 96.87 | 96.88 | 96.85 | 93.80 |
| R10 | 98.44 | 95.31 | 95.45 | 96.88 | 96.92 | 93.80 |
| Mean ± SD | 96.88 ± 1.85 | 96.72 ± 1.47 | 96.76 ± 1.39 | 96.80 ± 0.82 | 96.79 ± 0.84 | 93.65 ± 1.60 |
Fig. 10Error bar of validation performances
Validation performance (Unit: %)
| Model | CLFS | Sensitivity | Specificity | Precision | Accuracy | F1 | MCC |
|---|---|---|---|---|---|---|---|
| DenseNet-121 | I | 94.53 ± 2.34 | 94.69 ± 1.88 | 94.77 ± 1.67 | 94.61 ± 0.65 | 94.60 ± 0.70 | 89.32 ± 1.25 |
| II | 94.69 ± 1.59 | 94.84 ± 1.22 | 94.87 ± 1.22 | 94.77 ± 1.11 | 94.76 ± 1.13 | 89.56 ± 2.23 | |
| III | 95.00 ± 1.36 | 94.84 ± 1.41 | 94.90 ± 1.31 | 94.92 ± 0.87 | 94.93 ± 0.88 | 89.89 ± 1.74 | |
| IV | 95.00 ± 1.17 | 95.31 ± 1.56 | 95.35 ± 1.43 | 95.16 ± 0.77 | 95.14 ± 0.75 | 90.37 ± 1.52 | |
| DenseNet-169 | I | 95.47 ± 1.09 | 95.47 ± 2.36 | 95.58 ± 2.19 | 95.47 ± 0.91 | 95.48 ± 0.84 | 91.02 ± 1.81 |
| II | 95.47 ± 2.15 | 95.31 ± 0.70 | 95.38 ± 0.63 | 95.39 ± 1.02 | 95.39 ± 1.08 | 90.85 ± 2.03 | |
| III | 95.94 ± 2.23 | 95.63 ± 1.53 | 95.70 ± 1.47 | 95.78 ± 1.12 | 95.78 ± 1.15 | 91.64 ± 2.22 | |
| IV | 96.09 ± 1.44 | 96.25 ± 1.25 | 96.27 ± 1.25 | 96.17 ± 1.23 | 96.17 ± 1.24 | 92.37 ± 2.47 | |
| DenseNet-201 | I | 95.94 ± 1.59 | 95.78 ± 1.86 | 95.88 ± 1.73 | 95.86 ± 0.50 | 95.86 ± 0.49 | 91.80 ± 1.00 |
| II | 96.56 ± 1.53 | 96.09 ± 2.13 | 96.19 ± 1.90 | 96.33 ± 0.93 | 96.34 ± 0.89 | 92.73 ± 1.79 | |
| III | 96.41 ± 1.86 | 96.88 ± 1.21 | 96.90 ± 1.18 | 96.64 ± 1.21 | 96.63 ± 1.24 | 93.32 ± 2.42 | |
| IV | 96.88 ± 1.85 | 96.72 ± 1.47 | 96.76 ± 1.39 | 96.80 ± 0.82 | 96.79 ± 0.84 | 93.65 ± 1.60 |
Test performance based on best model “201-IV” found by proposed DenseNet-OTLS (unit: %)
| Run | Sen | Spc | Prc | Acc | F1 | MCC |
|---|---|---|---|---|---|---|
| R1 | 96.88 | 96.88 | 96.94 | 96.88 | 96.90 | 93.77 |
| R2 | 97.92 | 94.79 | 94.98 | 96.35 | 96.42 | 92.76 |
| R3 | 95.83 | 97.92 | 97.92 | 96.88 | 96.85 | 93.79 |
| R4 | 95.83 | 97.92 | 97.87 | 96.88 | 96.84 | 93.77 |
| R5 | 97.92 | 95.83 | 95.96 | 96.88 | 96.92 | 93.79 |
| R6 | 96.88 | 94.79 | 94.90 | 95.83 | 95.88 | 91.69 |
| R7 | 94.79 | 96.88 | 96.83 | 95.83 | 95.79 | 91.71 |
| R8 | 95.83 | 94.79 | 94.81 | 95.31 | 95.30 | 90.68 |
| R9 | 94.79 | 96.88 | 96.78 | 95.83 | 95.77 | 91.70 |
| R10 | 96.88 | 95.83 | 95.88 | 96.35 | 96.37 | 92.72 |
| Mean ± SD | 96.35± 1.07 | 96.25± 1.16 | 96.29± 1.11 | 96.30± 0.54 | 96.30± 0.56 | 92.64± 1.08 |
Fig. 11Comparison of validation performance and test performance
Comparison with COVID-19 identification approaches (Unit: %)
| Approach | Sen | Spc | Prc | Acc | F1 | MCC |
|---|---|---|---|---|---|---|
| RBFNN [ | 67.08 | 74.48 | 72.52 | 70.78 | 69.64 | 41.74 |
| K-ELM [ | 57.29 | 61.46 | 59.83 | 59.38 | 58.46 | 18.81 |
| ELM-BA [ | 57.08 ± 3.86 | 72.40 ± 3.03 | 67.48 ± 1.65 | 64.74 ± 1.26 | 61.75 ± 2.24 | 29.90 ± 2.45 |
| 6L-CNN [ | 81.04 ± 2.90 | 79.27 ± 2.21 | 79.70 ± 1.27 | 80.16 ± 0.85 | 80.31 ± 1.13 | 60.42 ± 1.73 |
| GoogLeNet [ | 76.88 ± 3.92 | 83.96 ± 2.29 | 82.84 ± 1.58 | 80.42 ± 1.40 | 79.65 ± 1.92 | 61.10 ± 2.62 |
| ResNet-18 [ | 78.96 ± 2.90 | 89.48 ± 1.64 | 88.30 ± 1.50 | 84.22 ± 1.23 | 83.31 ± 1.53 | 68.89 ± 2.33 |
| DenseNet-OTLS (Ours) | 96.35 ± 1.07 | 96.25 ± 1.16 | 96.29 ± 1.11 | 96.30 ± 0.54 | 96.30 ± 0.56 | 92.64 ± 1.08 |
Fig. 12Comparison of our method with seven state-of-the-art approaches
Composite learning factor versus simple learning factor (Unit: %)
| Setting | Sen | Spc | Prc | Acc | F1 | MCC |
|---|---|---|---|---|---|---|
| SLF | 94.37 ± 1.06 | 94.58 ± 1.60 | 94.62 ± 1.50 | 94.48 ± 0.78 | 94.48 ± 0.76 | 89.00 ± 1.56 |
| CLF(Ours) | 96.35 ± 1.07 | 96.25 ± 1.16 | 96.29 ± 1.11 | 96.30 ± 0.54 | 96.30 ± 0.56 | 92.64 ± 1.08 |
Fig. 13Error bar plot of CLF versus SFL
Effectiveness of proposed three-step preprocessing
| Setting | Sen | Spc | Prc | Acc | F1 | MCC |
|---|---|---|---|---|---|---|
| Only resizing | 93.33 ± 2.21 | 91.77 ± 2.75 | 91.94 ± 2.55 | 92.55 ± 2.07 | 92.61 ± 2.02 | 85.14 ± 4.13 |
| Preprocessing (Ours) | 96.35 ± 1.07 | 96.25 ± 1.16 | 96.29 ± 1.11 | 96.30 ± 0.54 | 96.30 ± 0.56 | 92.64 ± 1.08 |