| Literature DB >> 34847386 |
Abstract
Deep learning (DL) has shown great success in the field of medical image analysis. In the wake of the current pandemic situation of SARS-CoV-2, a few pioneering works based on DL have made significant progress in automated screening of COVID-19 disease from the chest X-ray (CXR) images. But these DL models have no inherent way of expressing uncertainty associated with the model's prediction, which is very important in medical image analysis. Therefore, in this paper, we develop an uncertainty-aware convolutional neural network model, named UA-ConvNet, for the automated detection of COVID-19 disease from CXR images, with an estimation of associated uncertainty in the model's predictions. The proposed approach utilizes the EfficientNet-B3 model and Monte Carlo (MC) dropout, where an EfficientNet-B3 model has been fine-tuned on the CXR images. During inference, MC dropout has been applied for M forward passes to obtain the posterior predictive distribution. After that mean and entropy have been calculated on the obtained predictive distribution to get the mean prediction and model uncertainty. The proposed method is evaluated on the three different datasets of chest X-ray images, namely the COVID19CXr, X-ray image, and Kaggle datasets. The proposed UA-ConvNet model achieves a G-mean of 98.02% (with a Confidence Interval (CI) of 97.99-98.07) and sensitivity of 98.15% for the multi-class classification task on the COVID19CXr dataset. For binary classification, the proposed model achieves a G-mean of 99.16% (with a CI of 98.81-99.19) and a sensitivity of 99.30% on the X-ray Image dataset. Our proposed approach shows its superiority over the existing methods for diagnosing the COVID-19 cases from the CXR images.Entities:
Keywords: COVID-19 automatic screening uncertainty estimation Monte Carlo dropout pre-trained EfficientNet CNN Chest X-ray images
Year: 2021 PMID: 34847386 PMCID: PMC8609674 DOI: 10.1016/j.compbiomed.2021.105047
Source DB: PubMed Journal: Comput Biol Med ISSN: 0010-4825 Impact factor: 4.589
Fig. 1Examples of (a) COVID-19 and (b) Viral Pneumonia in X-ray images as shown in the top and the bottom row, respectively. Rectangle markers represent findings (disease patterns) in the CXR images.
Details of the original top layers and new top layers.
| Original top layer (s) | New top layer (s) | ||||
|---|---|---|---|---|---|
| Layer | Output size | Trainable | Layer | Output size | Trainable |
| Global average | 1536 | 0 | Flatten | 75264 | 0 |
| Dropout | 1536 | 0 | Dense | 1024 | 77071360 |
| Dense | 1000 | 1537000 | MC dropout(0.425) | 1024 | 0 |
| Activation | 1000 | 0 | Softmax | 3 | 3075 |
Fig. 2Schematic diagram of the proposed method. Top layer of EfficientNet-B3 is replaced with Flatten, Dense, MC dropout and Softmax layers. During the inference test images are given to the trained model for M forward passes and obtained predictive distribution of size M.
Details of the Dataset-A.
| Classes | Number of Images | Number of Patients |
|---|---|---|
| COVID-19 | 546 | 332 |
| NORMAL | 1139 | 1015 |
| PNEOMONIA | 1355 | 583 |
Fig. 35-fold cross-validation strategy.
Details of the hyper-parameter settings.
| CNN (s) | Hyper-parameters | |||
|---|---|---|---|---|
| Optimizer | Learning Rate | Batch Size | Epochs | |
| VGG19 | RMSprob | 0.0001 | 16 | 30 |
| ResNet152 | SGD | 0.001 | 8 | 30 |
| Xception | RMSprob | 0.0001 | 16 | 30 |
| DenseNet169 | RMSprob | 0.0001 | 16 | 30 |
| MobileNet | SGD | 0.001 | 8 | 20 |
| NASNet-Large | SGD | 0.001 | 8 | 20 |
| Inception-V3 | SGD | 0.001 | 8 | 30 |
| EfficientNet (B0–B5) | RMSprob | 0.0001 | 32 | 20 |
| UA-ConvNet | RMSprob | 0.0001 | 32 | 20 |
Performance of the UA-ConvNet model for multi-class classification task on the Dataset-A.
| Metrics | Fold1 | Fold2 | Fold3 | Fold4 | Fold5 | Mean |
|---|---|---|---|---|---|---|
| 98.49 | 96.82 | 96.74 | 98.60 | 98.68 | ||
| 98.26 | 97.16 | 97.46 | 98.76 | 99.13 | ||
| 99.00 | 98.21 | 98.41 | 99.12 | 99.42 | ||
| 98.36 | 96.99 | 97.04 | 98.66 | 98.89 | ||
| 98.09 | 96.65 | 96.83 | 97.83 | 98.93 | ||
| 98.15 | 97.07 | 97.35 | 98.34 | 99.18 | ||
| 97.14 | 94.78 | 95.14 | 97.43 | 98.18 | ||
| 99.47 | 99.65 | 99.40 | 99.84 | 99.89 |
“Prec”: Precision, “Sens”: Sensitivity, “Spec”: Specificity.
“F1-Sc”: F1-Score, “Acc”: Accuracy.
Performance of the UA-ConvNet model for binary classification task on the Dataset-B.
| Metrics | Fold1 | Fold2 | Fold3 | Fold4 | Fold5 | Mean |
|---|---|---|---|---|---|---|
| 99.51 | 98.08 | 96.30 | 100 | 100 | ||
| 98.00 | 99.50 | 99.00 | 100 | 100 | ||
| 98.00 | 99.40 | 99.00 | 100 | 100 | ||
| 98.73 | 98.77 | 97.57 | 100 | 100 | ||
| 98.90 | 99.20 | 98.69 | 100 | 100 | ||
| 97.15 | 99.48 | 99.18 | 100 | 100 | ||
| 97.50 | 97.57 | 95.26 | 100 | 100 | ||
| 99.97 | 99.99 | 99.95 | 100 | 100 |
Confusion Matrices for multi-class classification on the Dataset-A.
| Fold1 | Fold2 | Fold3 | Fold4 | Fold5 | ||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Predicted | Predicted | Predicted | Predicted | Predicted | ||||||||||||
| C | N | P | C | N | P | C | N | P | C | N | P | C | N | P | ||
| C | 107 | 0 | 2 | 110 | 0 | 1 | 108 | 0 | 1 | 110 | 0 | 0 | 108 | 0 | 0 | |
| N | 0 | 227 | 1 | 0 | 219 | 8 | 1 | 227 | 0 | 0 | 227 | 1 | 0 | 228 | 0 | |
| P | 0 | 8 | 263 | 3 | 8 | 258 | 3 | 14 | 254 | 0 | 9 | 265 | 2 | 5 | 260 | |
“C”: COVID19, “N”: Normal, “N”: Pneumonia.
Fig. 4ROC curve for multi-class classification task on the different folds of Dataset-A.
Fig. 5ROC curve for binary classification task on the different folds of Dataset-B.
Fig. 6Model classification error for different number of forward passes (or sample size (k)); where (a) represent classification error on the Dataset-A for multi-class classification task, and (b) represent classification error on the Dataset-B for binary classification task.
Fig. 7UA-ConvNet model posterior predictive distribution for M forward passes corresponding to different folds of Dataset-A.
G-mean, standard deviation, Entropy and confidence interval for multi-class classification task on Dataset-A.
| Fold (s) | G-mean (%) | CI(%) @ 95% | SD | Entropy |
|---|---|---|---|---|
| Fold1 | 98.15 | [98.13–98.17] | 0.0014 | 0.0054 |
| Fold2 | 97.07 | [97.04–97.11] | 0.0023 | 0.0327 |
| Fold3 | 97.35 | [97.32–97.38] | 0.0022 | 0.0281 |
| Fold4 | 98.34 | [98.31–98.36] | 0.0020 | 0.0326 |
| Fold5 | 99.18 | [99.17–99.20] | 0.0013 | 0.0089 |
G-mean, standard deviation, Entropy and confidence interval binary classification task on Dataset-B.
| Fold(s) | G-mean (%) | CI(%) @ 95% | SD | Entropy |
|---|---|---|---|---|
| Fold1 | 97.15 | [96.27–98.04] | 0.0101 | 0.0075 |
| Fold2 | 99.48 | [99.44–99.51] | 0.0025 | 0.0029 |
| Fold3 | 99.18 | [99.15–99.21] | 0.0024 | 0.0054 |
| Fold4 | 100 | [100−100] | 0.0000 | 0.0000 |
| Fold5 | 100 | [100−100] | 0.0000 | 0.0000 |
Fig. 8X-ray images and their predicted labels, with class probabilities (COVID-19: P_c, Normal: P_n, Pneumonia: P_p) and uncertainty produced by UA-ConvNet model. Where, first, middle and last columns represent images with low confidence (or high uncertainty (E)), average confidence, and high confidence, respectively.
Portability performance of the proposed UA-ConvNet model on Dataset-A and Dataset-C.
| Prec | Sens | Spec | F1-Sc | MCC | G-mean | CI(%) @ | SD | Entropy | |
|---|---|---|---|---|---|---|---|---|---|
| Dataset-C ⇒Dataset-C | 99.04 | 98.27 | 99.17 | 98.64 | 97.64 | 97.83 | 97.75–97.91 | 0.0040 | 0.0299 |
| Dataset-C ⇒Dataset-A | 97.42 | 95.30 | 98.11 | 96.24 | 94.78 | 94.85 | 94.78–94.92 | 0.0036 | 0.0212 |
| Dataset-A ⇒Dataset-A | 98.49 | 98.26 | 99.00 | 98.36 | 97.14 | 98.15 | 98.13–98.17 | 0.0014 | 0.0054 |
| Dataset-A ⇒Dataset-C | 97.63 | 96.01 | 97.89 | 96.74 | 94.07 | 96.17 | 96.11–96.23 | 0.0030 | 0.0097 |
“Dataset-A ⇒Dataset-C”: UA-ConvNet model has been trained on Dataset-A and tested on Dataset-C.
Fig. 9Performances of pre-trained CNNs on the Dataset-A and Dataset-B.
Fig. 10Performances of EfficientNet models on the Dataset-A and Dataset-B.
Performance comparison with the existing methods.
| Author (s) | Method | Dataset∖Subjects | Accuracy (%) |
|---|---|---|---|
| Narin et al. [ | ResNet50, InceptionV3, | COVID-19: 50, | Binary: 98 |
| Zhang et al. [ | Two-stage transfer: | COVID-19: 189, | Multi-class: 91.08 |
| Waheed et al. [ | Auxiliary Classifier Generative | COVID-19: 403, | Binary: 95 |
| Togacar et al. [ | Fuzzy color, Stacking, | COVID-19: 295, | Multi-class: 99.27 |
| Oh et al. [ | ResNet18 | Normal:191, | Multi-class: 91.9 |
| Gianchandani et al. [ | Ensemble models: | COVID-19: 423, | Binary: 99.21 |
| Wang et al. [ | COVID-Net | COVID-19: 358, | Multi-class: 93.3 |
| Ozturk et al. [ | DarkCovidNet: | Binary: 98.08 | |
| Maheshwari et al. [ | LBP, image-based features | Multi-class: 96.99 | |
| Gour et al. [ | Stacked CNN model: | Multi-class: 92.74 | |
| UA-ConvNet: | Multi-class: | ||
| Multi-class: | |||
| Binary: |