| Literature DB >> 33100403 |
Mohammad Shorfuzzaman1, M Shamim Hossain2,3.
Abstract
Various AI functionalities such as pattern recognition and prediction can effectively be used to diagnose (recognize) and predict coronavirus disease 2019 (COVID-19) infections and propose timely response (remedial action) to minimize the spread and impact of the virus. Motivated by this, an AI system based on deep meta learning has been proposed in this research to accelerate analysis of chest X-ray (CXR) images in automatic detection of COVID-19 cases. We present a synergistic approach to integrate contrastive learning with a fine-tuned pre-trained ConvNet encoder to capture unbiased feature representations and leverage a Siamese network for final classification of COVID-19 cases. We validate the effectiveness of our proposed model using two publicly available datasets comprising images from normal, COVID-19 and other pneumonia infected categories. Our model achieves 95.6% accuracy and AUC of 0.97 in diagnosing COVID-19 from CXR images even with a limited number of training samples.Entities:
Keywords: COVID-19 diagnosis; CXR images; Contrastive loss; Multi-shot learning; Siamese network
Year: 2020 PMID: 33100403 PMCID: PMC7568501 DOI: 10.1016/j.patcog.2020.107700
Source DB: PubMed Journal: Pattern Recognit ISSN: 0031-3203 Impact factor: 7.740
Fig. 1COVID-19 trend in global scale. Graph shows total number of confirmed, active, death, and recovered cases. (For interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article.)
Fig. 2An example of a K-way N-shot learning problem where K = 3 and N = 2 in the support set. Query set images need to be classified from 3 available classes {normal, COVID-19 positive, and non-COVID pneumonia}.
Fig. 3High-level architecture of deep Siamese neural network for n-shot COVID-19 classification. (Zooming may be required for superior view). (For interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article.)
Fig. 4Contrastive loss showing the margin m. The blue solid line signifies the loss function for the dissimilar pairs and the dotted red line refers to the same for similar pairs [33]. (For interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article.)
Fig. 5An example training strategy for 2-shot, 3-class image classification task.
Training algorithm for k-way n-shot learning.
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| Update parameter |
| |
Dataset split statistics.
| Class | Pre-training of VGG-16 encoder network | Siamese network ( | ||
|---|---|---|---|---|
| Training | Testing | Training | Testing | |
| Normal | 160 | 66 | 10 | 216 |
| Non-COVID pneumonia | 160 | 66 | 10 | 216 |
| COVID-19 | 160 | 66 | 10 | 216 |
| Total | 480 | 198 | 30 | 648 |
Performance results for various n-shot settings with contrastive loss. 3-way represents 3-class labels.
| Model | Accuracy | Precision | Recall | Specificity | F1-score | AUC |
|---|---|---|---|---|---|---|
| MetaCOVID (3-way, 7-shot) | 0.925 | 0.945 | 0.936 | 0.953 | 0.940 | 0.955 |
| MetaCOVID (3-way, 8-shot) | 0.936 | 0.951 | 0.945 | 0.965 | 0.938 | 0.962 |
| MetaCOVID (3-way, 9-shot) | 0.948 | 0.966 | 0.955 | 0.975 | 0.947 | 0.974 |
| MetaCOVID 3-way, 10-shot) | 0.956 | 0.970 | 0.960 | 0.980 | 0.965 | 0.975 |
Performance results for various 3-way, n-shot settings with cross entropy loss.
| Model | Accuracy | Precision | Recall | Specificity | F1-score | AUC |
|---|---|---|---|---|---|---|
| MetaCOVID (3-way, 7-shot) | 0.890 | 0.927 | 0.915 | 0.935 | 0.916 | 0.933 |
| MetaCOVID (3-way, 8-shot) | 0.915 | 0.935 | 0.919 | 0.940 | 0.922 | 0.948 |
| MetaCOVID (3-way, 9-shot) | 0.923 | 0.938 | 0.939 | 0.948 | 0.938 | 0.954 |
| MetaCOVID 3-way, 10-shot) | 0.938 | 0.949 | 0.953 | 0.964 | 0.950 | 0.957 |
Fig. 6Training and validation accuracy and loss for 3-way, 10-shot learning settings with (a) contrastive loss (b) cross-entropy loss.
Performance comparison between the proposed Siamese network model (with 3-way, 10-shot learning) and other pre-trained CNN models.
| Model | Acc. | Precision | Recall | Specificity | F1-score | AUC |
|---|---|---|---|---|---|---|
| InceptionV3 | 0.875 | 0.826 | 0.950 | 0.800 | 0.883 | 0.900 |
| Xception | 0.955 | 0.977 | 0.956 | 0.988 | 0.966 | 0.980 |
| InceptionResNetV2 | 0.900 | 0.833 | 1.00 | 0.800 | 0.908 | 0.900 |
| VGG-16 | 0.933 | 0.956 | 0.956 | 0.976 | 0.956 | 0.954 |
| MetaCOVID (3-way, 10-shot) | 0.956 | 0.970 | 0.960 | 0.980 | 0.965 | 0.975 |
Performance results of our model with contrastive loss for various 2-way, n-shot settings for 2-class (normal, COVID-19) classification.
| Model | Accuracy | Precision | Recall | Specificity | F1-score | AUC |
|---|---|---|---|---|---|---|
| MetaCOVID (2-way, 7-shot) | 0.940 | 0.955 | 0.945 | 0.958 | 0.949 | 0.965 |
| MetaCOVID (2-way, 8-shot) | 0.948 | 0.963 | 0.955 | 0.975 | 0.958 | 0.975 |
| MetaCOVID (2-way, 9-shot) | 0.950 | 0.975 | 0.965 | 0.980 | 0.969 | 0.982 |
| MetaCOVID 2-way, 10-shot) | 0.965 | 0.980 | 0.970 | 0.984 | 0.974 | 0.989 |