| Literature DB >> 34769600 |
Debaditya Shome1, T Kar1, Sachi Nandan Mohanty2, Prayag Tiwari3, Khan Muhammad4, Abdullah AlTameem5, Yazhou Zhang6, Abdul Khader Jilani Saudagar5.
Abstract
In the recent pandemic, accurate and rapid testing of patients remained a critical task in the diagnosis and control of COVID-19 disease spread in the healthcare industry. Because of the sudden increase in cases, most countries have faced scarcity and a low rate of testing. Chest X-rays have been shown in the literature to be a potential source of testing for COVID-19 patients, but manually checking X-ray reports is time-consuming and error-prone. Considering these limitations and the advancements in data science, we proposed a Vision Transformer-based deep learning pipeline for COVID-19 detection from chest X-ray-based imaging. Due to the lack of large data sets, we collected data from three open-source data sets of chest X-ray images and aggregated them to form a 30 K image data set, which is the largest publicly available collection of chest X-ray images in this domain to our knowledge. Our proposed transformer model effectively differentiates COVID-19 from normal chest X-rays with an accuracy of 98% along with an AUC score of 99% in the binary classification task. It distinguishes COVID-19, normal, and pneumonia patient's X-rays with an accuracy of 92% and AUC score of 98% in the Multi-class classification task. For evaluation on our data set, we fine-tuned some of the widely used models in literature, namely, EfficientNetB0, InceptionV3, Resnet50, MobileNetV3, Xception, and DenseNet-121, as baselines. Our proposed transformer model outperformed them in terms of all metrics. In addition, a Grad-CAM based visualization is created which makes our approach interpretable by radiologists and can be used to monitor the progression of the disease in the affected lungs, assisting healthcare.Entities:
Keywords: COVID-19; data science; deep learning; grad-CAM; healthcare; interpretability; transfer learning; vision transformer
Mesh:
Year: 2021 PMID: 34769600 PMCID: PMC8583247 DOI: 10.3390/ijerph182111086
Source DB: PubMed Journal: Int J Environ Res Public Health ISSN: 1660-4601 Impact factor: 3.390
Figure 1The proposed ViT model for COVID-19 detection.
Figure 2Custom MLP block attached to the Vision Transformer.
Figure 3Performance measures during training.
Figure 4Data set distribution and description.
Figure 5Confusion matrix for both types of classification.
Evaluation of the proposed model.
| Model | Accuracy | Precision | Recall | F1 Score | AUC |
|---|---|---|---|---|---|
|
| 0.98 | 0.97 | 0.97 | 0.97 | 0.99 |
|
| 0.92 | 0.93 | 0.89 | 0.91 | 0.98 |
Performance comparison of our COVID-Transformer with baseline models on the multi-class classification problem.
| Model | Accuracy | Precision | Recall | F1 Score | AUC |
|---|---|---|---|---|---|
| 0.90 | 0.89 | 0.91 | 0.89 | 0.92 | |
| 0.89 | 0.88 | 0.89 | 0.88 | 0.92 | |
| 0.90 | 0.90 | 0.89 | 0.90 | 0.92 | |
| 0.88 | 0.87 | 0.86 | 0.86 | 0.93 | |
| 0.87 | 0.87 | 0.85 | 0.86 | 0.90 | |
| 0.90 | 0.92 | 0.87 | 0.90 | 0.93 | |
| 0.88 | 0.90 | 0.85 | 0.87 | 0.92 | |
|
|
|
|
|
|
|
Figure 6Grad-CAM visualization for the three classes.
Figure 7Flow of deployable solution.