Literature DB >> 35494338

COVID-19 prognosis using limited chest X-ray images.

Arnab Kumar Mondal1.   

Abstract

The COrona VIrus Disease 2019 (COVID-19) pandemic is an ongoing global pandemic that has claimed millions of lives till date. Detecting COVID-19 and isolating affected patients at an early stage is crucial to contain its rapid spread. Although accurate, the primary viral test 'Reverse Transcription Polymerase Chain Reaction' (RT-PCR) for COVID-19 diagnosis has an elaborate test kit, and the turnaround time is high. This has motivated the research community to develop CXR based automated COVID-19 diagnostic methodologies. However, COVID-19 being a novel disease, there is no annotated large-scale CXR dataset for this particular disease. To address the issue of limited data, we propose to exploit a large-scale CXR dataset collected in the pre-COVID era and train a deep neural network in a self-supervised fashion to extract CXR specific features. Further, we compute attention maps between the global and the local features of the backbone convolutional network while finetuning using a limited COVID-19 CXR dataset. We empirically demonstrate the effectiveness of the proposed method. We provide a thorough ablation study to understand the effect of each proposed component. Finally, we provide visualizations highlighting the critical patches instrumental to the predictive decision made by our model. These saliency maps are not only a stepping stone towards explainable AI but also aids radiologists in localizing the infected area.
© 2022 Elsevier B.V. All rights reserved.

Entities:  

Keywords:  AI for COVID-19; COVID-19 detection from limited data; COVID-19 detection using CXR; Chest radiography; Deep learning; Self-supervised learning

Year:  2022        PMID: 35494338      PMCID: PMC9035620          DOI: 10.1016/j.asoc.2022.108867

Source DB:  PubMed          Journal:  Appl Soft Comput        ISSN: 1568-4946            Impact factor:   8.263


Introduction

The COrona VIrus Disease 2019 (COVID-19), epi-centered in Hubei Province of the People’s Republic of China, spread so rapidly across the globe that the World Health Organization (WHO) declared COVID-19 a Public Health Emergency of International Concern on 30 January 2020, and finally a pandemic on 11 March 2020 [1]. It has caused a massive threat to global health with 174,918,667 cases of confirmed coronavirus and 3,782,490 deaths as of 12 June 2021. Once infected with COVID-19, one may experience fever, cough, and respiratory illness. Some may also experience shortness of breath, muscle or body aches, headache, loss of taste or smell, sore throat, and diarrhea [2], [3]. The virus can cause pneumonia or breathing problems in severe cases, leading to multi-organ failure and death [4]. Due to the exponential growth of COVID-19 patients, there is a shortage in supply of diagnostic kits, a limited number of beds in the hospitals to care for critical patients, a dearth of ventilators, scarcity in personal protective equipment (PPE) for healthcare personnel. Despite various preventive measures (such as complete lockdown) adopted by the government of different countries to contain the disease and delay the spread, several developed countries have faced a critical care crisis, and the health system has come to the verge of collapse. It is, therefore, of utmost importance to screen the positive COVID-19 patients accurately for efficient utilization of limited resources. Reverse Transcription Polymerase Chain Reaction (RT-PCR) [5], [6] is the most preferred viral test for COVID-19 detection due to its high sensitivity and specificity. However, the turn-around time of RT-PCR is high. Consequently, chest radiography such as computerized tomography (CT) scan and X-ray imaging-based detection techniques have emerged as an alternative modality for screening COVID-19 patients. With these modalities, researchers have observed that COVID-19 patients’ lungs exhibit ground-glass opacity and/or mixed ground-glass opacity and mixed consolidation that can separate COVID-19-positive cases from COVID-19-negative cases [7], [8]. In contrast to conventional diagnostic methods, X-ray offers several advantages as it is fast, can simultaneously analyze numerous cases, inexpensive and widely available. It can be very useful in hospitals with limited testing kits and resources. Deep Machine Learning has revolutionized the field of health care by accurately analyzing, identifying, and classifying patterns in medical images [9]. Artificial neural networks are able to diagnose a variety of illnesses with a high degree of accuracy. The reason for such success is that deep learning techniques do not rely on manual handcrafted features but rather learn features automatically from data itself. This allows the algorithm to be applicable on a broader variety of use cases than traditional machine learning methods and is also faster and more accurate in many cases. Motivated by the remarkable performance of CheXNet in Pneumonia detection from chest X-ray images, artificial intelligence (AI) researchers have put a lot of effort into designing machine learning (ML) algorithms for automated detection of COVID-19 using chest X-rays. However, the biggest challenge lies in the fact that COVID-19 being a novel disease, a limited number of sample images are available for training deep neural networks. Motivated by this, in this work, we propose a novel framework that can be trained using limited labeled data for COVID-19 detection using chest X-rays. In this work, our contributions are as follows. We adopt a self-supervised training methodology to train a CXR feature extractor (a convolutional backbone network) on a large-scale chest X-ray dataset. We design a local–global-attention-based classification network consisting of the pre-trained feature extractor, an attention block, and a classification head. We empirically demonstrate the effectiveness of the proposed framework in the low data regime through extensive experimentation and ablation studies. We present clinically interpretable saliency maps, which are helpful for disease localization and patient triage. The remainder of this paper is structured as follows: Section 2 provides an overview of related work; Section 3 describes the procedural and methodological stages of the development of this solution; Section 4 evaluates the proposed method and assesses the predictions; finally, Section 5 critically discusses the advantages and the limitations of the proposed framework.

Related work

Several deep neural frameworks [10], [11], [12], [13], [14], [15] have been proposed in the past to identify different thoracic diseases such as Pneumonia using chest X-ray (CXR) images and surpassed average radiologist performance. ChestX-ray8 [11] (later extended to constitute ChestX-ray14 dataset), and CheXpert [16] are two large-scale datasets of chest X-rays (CXR) that facilitate the training of deep neural networks (DNN) for automating the interpretation of a wide variety of thoracic diseases. ChexNet [16] is a deep neural network, built using DenseNet-121 [17], for Pneumonia detection using chest X-ray images and it achieved excellent results surpassing average radiologist performance. ChestNet [12] is another deep neural network for thoracic diseases diagnosis using chest radiography images. The authors in [14] propose to learn channel-wise, element-wise, and scale-wise attention (triple attention) simultaneously to classify 14 thoracic diseases using chest radiography. Thorax-Net [15] is an attention regularized deep neural network for the classification of thoracic diseases on chest radiography. Motivated by this, the research community has examined the possibility of COVID-19 prognosis using CXR.

Traditional machine learning for COVID-19 detection using CXR

The proposal in [18] leverages an enhanced cuckoo search algorithm to determine the most significant CXR features and train a -nearest neighbor (KNN) classifier to distinguish between COVID-19 positive and negative cases. In this work, features were extracted from X-ray images using standard feature extraction techniques such as Fractional Zernike Moments (FrZMs), Wavelet Transform (WT), Gabor Wavelet Transform (GW), and Gray Level Co-Occurrence Matrix (GLCM), followed by a fractional order cuckoo search method where the levy flight distribution was replaced with better suited heavy tailed distributions for selecting the most relevant features. Following feature selection, a KNN was used for classification. The work in [19] employs a new set of descriptors, Fractional Multichannel Exponent Moments (FrMEMs) to extract orthogonal moment features. Next, Manta Ray Foraging Optimization (MRFO) using Differential evolution (DE) is utilized to select the most relevant features. Finally, a -nearest neighbor (KNN) classifier is used for prediction. A novel shape-dependent Fibonacci-p patterns-based feature descriptor is proposed in [20] for CXR features extraction, which are classified using conventional ML algorithms such as support vector machine (SVM), -nearest neighbor (KNN), Random Forest, AdaBoost, Gradient Tree Boosting, and Decision Trees. In [21], the author uses Histogram of Oriented Gradients (HOG), Gray-Level Co-Occurrence Matrix (GLCM), Scale-Invariant Feature Transform (SIFT), and Local Binary Pattern (LBP) methods in the feature extraction phase. Next, Principle Component Analysis (PCA) is applied for feature selection. Finally, k-NN, SVM, Bag of Tree, and Kernel Extreme Learning Machine (K-ELM) are used for final classification.

Deep learning for COVID-19 detection using CXR

Many of the existing deep learning methods [22], [23], [24], [25], [26], [27], [28], [29] use the transfer learning approach by finetuning pre-trained networks such as ResNet-18 [30] or ResNet-50 [30], DenseNet-121 [17], InceptionV3 [31], Xception [32], etc., on COVID- 19 CXR datasets. COVID-SDNet [28] combines segmentation, data-augmentation and data transformations together with a ResNet-50 [30] for inference. The authors in [28] define a novel three-stage segmentation-classification pipeline to solve a binary classification task between COVID-19 and non-COVID-19 CXR. First, the lung region is cropped from CXR using bounding box segmentation. Next, a GAN based class-inherent transformation network is employed to generate two class inherent transformations and from each input image . Finally, the transformed images are used to solve a four-class classification problem using CNN with a Resnet-50 [30] backbone and an aggregation strategy is designed in order to obtain the final class. As the number of classes increase, so will the number of generators to be trained in stage two, which makes scaling inefficient for multi-class classification. In [24] an ensemble of off the shelf pretrained CNNs – InceptionV3 [31], MobileNetV2 [33], ResNet101 [30], NASNet [34] and Xception [32] – is first fine tuned on the chest xray dataset. Their final layer representations are then stacked and then passed through a MLP for COVID-19 diagnosis. The Xception [32] backbone is used in CoroNet [22] for extracting CXR features which are classified using the MLP classification head. In [35] proposed DeepCoroNet where the CXR images are pre-processed using a sobel filter followed by marker-controlled watershed segmentation and then a deep LSTM network is used for classification. The work in [36] uses Google’s Big Transfer models with DenseNet, InceptionV3 and Inception-ResNetV4 models for COVID-19 classification using chest X-rays. COVID-Net [37] proposes a custom architecture for CXR-based COVID-19 detection using a human–machine collaborative design strategy. However, limited COVID-19 samples restrict the generalizability of such large-capacity models. To address this issue, MAG-SD [38] employs a multi-scale attention-guided deep network to augment the data and formulates a new regularization term utilizing soft distance between predictions, to regularize the classifier from producing contradicted output for one target. An attention-based teacher–student framework is proposed in [39]. The teacher network extracts global features and focuses on the infected regions to generate attention maps. An image fusion module transfers the attention knowledge to the student network. CHP-Net [40] involves a discrimination network for lung feature extraction to discriminate COVID-19 cases and a localization network to localize and assign the recognized X-ray images into the left lung, right lung or bipulmonary. In [41] a federated learning model is developed while keeping in mind the privacy of the patients. Individual hospitals or care centers are considered as nodes which have their own datasets and share a common diagnosis model provided by a central server. The individual nodes update the model according to the dataset that they have and their updated weights are averaged and the common server model is updated. In [42] a multimodal system is developed based on data consisting of breathing sounds and chest X-ray images. Sound data is converted to spectrograms and convolutional neural networks are used for analysis for both sound data and chest xray images. An InceptionV3 network is used followed by an MLP for COVID-19 diagnosis. The authors in [43] propose a convolutional CapsNet for COVID-19 detection from chest X-ray images in binary as well as multi-class classification settings. xViTCOS [44] propose a vision transformer based deep neural classifier for COVID-19 prognosis. Illustration of our proposed framework for COVID-19 detection using limited chest X-ray images.

Proposed method

Supervised learning usually demands a large amount of labeled data. However, collecting quality annotated data is expensive, especially for medical applications. Moreover, COVID-19 being a novel disease, there is a scarcity of well-curated high volume datasets. Therefore, we propose to utilize a self-supervised training methodology to address this issue of data scarcity. In the first stage, we train a convolutional neural network on a large-scale CXR dataset, CheXpert [16] for extracting robust CXR features with self-supervision. Next, we utilize limited COVID-19 CXR images to train a classification network that uses the pretrained backbone to extract local and global features, computes attention maps, and predicts the class label.

Self-supervised pretraining for representation learning

The fundamental concept behind self-supervised learning is to design some auxiliary pre-text tasks such that the model discovers the underlying structure of the data while solving those tasks. Several state-of-the-art self-supervised methods [45], [46], [47], [48] rely on contrastive strategy to induce similarity between positive pairs (different augmented views of the same image) and dissimilarity between negative pairs (augmented views from different images). These methods, however, require either large batch size, memory bank, or custom mining strategies while selecting negative pairs. Bootstrap Your Own Latent (BYOL) [49] mitigates this issue associated with negative pair selection. In this work, we propose to use BYOL for representation learning. As illustrated in Fig. 1, BYOL consists of two neural networks, viz., online and target networks. These two networks interact and learn together. The online network consists of three sub-networks: an encoder , a projector , and a predictor . denotes the set of trainable parameters of the online network. To break the symmetry between the online and target pipeline, the target network is comprised of two sub-networks: an encoder and a projector . The parameters, of the target network are slow moving average of the online network parameters, i.e., where denotes the target decay rate.
Fig. 1

Illustration of our proposed framework for COVID-19 detection using limited chest X-ray images.

At the beginning of each training step, an original image, is drawn uniformly from the CheXpert [16] dataset. Next, two sets of randomly chosen transformations, are applied on the original image, to obtain two distinct augmented views, of the underlying true image. During training, is fed into the online network, and is fed into the target network. The online network generates a representation, , a projection, , and a prediction, . The target network produces a target representation, and a target projection, . Since the target network is derived from the online network, the online representations should be predictive of the target representations. Consequently, BYOL is trained to maximize the similarity between these two representations. Mathematically, the online network is trained to minimize the mean squared error between the normalized online prediction, and the normalized target projection, : To make the loss symmetric, next, is passed through the target network and is passed through the online network and loss is computed according to Eq. (2). The total loss is now given as, The sub-script, in implies that, only the online network is updated to minimize , and the target network is updated as the exponential moving average as indicated in Eq. (1).

Multi-scale spatial attention based classifier

In the second stage of our proposed method, we utilize the pretrained backbone from the previous step and design a spatial attention network based on the local and the global features. Attention mechanism are widely adapted to enhance the performance of deep neural networks on various downstream tasks such as machine translation, text generation in natural language processing and object classification, image captioning, inpainting, etc., in computer vision. Attention in computer vision tasks can broadly be categorized into spatial attention [50], [51] that captures the local context and channel attention [52] that captures the global semantics. Several works [53], [54] consider a combination of both channel-wise and spatial attention. In this work, we adopt the soft trainable visual attention proposed in [55]. Fig. 2 presents an overview of the attention mechanism. We extract local and global features using pretrained feature extractor backbone. ‘Local features’ refer to features extracted by some convolutional layer of the backbone network that have a limited receptive field. In other words the receptive field is a contiguous proper subset of the image (local). The contents of the ‘local features’ can be more specific to a certain region on the image, while ‘global features’ use the entire image as their information source. We insert three attention estimators after ‘layer2’, ‘layer3’, and ‘layer4’ (layer names as per PyTorch implementation) to capture coarse-to-fine attention maps at multiple levels. The local features extracted at these three layers together with the global feature at the penultimate ‘avgpool’ layer produce three attended encodings, which are concatenated and fed into a final classification head.
Fig. 2

Illustration of the attention mechanism.

Let, denote the set of feature vectors extracted at a given convolutional layer . Where is the vector of output activations at the spatial location of total spatial locations in the layer. The global feature vector, , has the entire input image as the receptive field. Let be the compatibility function that computes a scalar compatibility score between two vectors of equal dimension. Since the dimensionality of the local features and the global features do not match, we first project the low-dimensional local features to the high-dimensional space of . Next, the compatibility score function is employed to compute the compatibility scores as follows: where and is a learnable vector. Illustration of the attention mechanism. The compatibility scores are then normalized using the softmax function to compute the attention maps. The attended representations are finally computed as follows: In this work, we concatenate the three representations obtained from three intermediate layers into a single vector, and feed to a linear classification head.

Experiments

In this section, we describe the dataset used in this work and discuss the experimental results.

Dataset

While some of the works [56] evaluate their proposed algorithm using private datasets, many other works [22], [37], [57] resort to publicly available datasets. In this work, we combine data from several publicly available repositories to create a custom dataset with four classes: Normal, Bacterial Pneumonia, Viral Pneumonia (non-COVID-19), COVID-19. As in [22], we collected Normal, Bacterial Pneumonia, and non COVID-19 Viral Pneumonia chest X-ray images from the Kaggle repository ‘Chest X-ray Images (Pneumonia)’ [58], which is derived from [59]. Chest X-ray images of COVID-19 patients were obtained from the Kaggle repository ‘COVIDx CXR-2’ [60], which is a combination of several publicly available resources [61], [62], [63], [64], [65], [66]. ‘COVIDx CXR-2’ [60] specifies only train-test split of the dataset. We hold out 20% training examples for automatic model selection based on its performance over the validation set. The validation set in the standard split of ‘Chest X-ray Images (Pneumonia)’ [58] dataset contains only 8 images per class. To avoid a huge class imbalance in the validation set, we combine the training and validation examples and split them into an 80:20 ratio. Table 1 tabulates a summarized description of split-wise image distribution. Note that the test split in the standard data division is left untouched to ensure there is no patient-wise information leakage as multiple CXR images of a patient might be present in the dataset.
Table 1

Summarized description of CXR dataset.

SplitNormalnon-COVID Pneumonia
COVID-19Total
BacterialViral
Train10792030107617265911
Validation2705082694321479
Test234242148200824
Summarized description of CXR dataset. Comparison of performance of the proposed method on chest X-ray dataset against state-of-the-art methods.

Implementation details

Image preprocessing and augmentation

In our compiled dataset and the CheXpert [16] dataset, the images are of variable sizes. To address this issue, we resize all the images to a fixed size of 256 × 256. For training BYOL, we randomly choose an image from the CheXpert dataset, select a random patch and resize it to 224 ×224. Next, the image is flipped randomly horizontally with 0.5 probability. Apart from these spatial/geometric transformations of data, we apply appearance transformations on the image. Specifically, we apply a random color distortion transformation consisting of a random sequence of brightness, contrast, saturation, and hue adjustments [67], [68]. As noted in previous work [45], stronger color jittering helps self-supervised algorithms learn better representation. We utilize PyTorch’s standard implementation (torchvision.transforms.ColorJitter) for performing color distortion. Following [45], we set brightness, contrast and saturation jitter factor uniformly from . The hue jitter factor is chosen uniformly from . Color distortion is applied randomly 80% of the time. Finally, random Gaussian blur is applied to the patches, and the patches are normalized. We blur the image with 0.5 probability using a Gaussian kernel. We randomly sample , and the kernel size is set to be . In the second stage of training, we randomly choose an image from the compiled CXR dataset, select a random patch and resize it to 224 × 224 with a random horizontal flip. Finally, the patches are normalized before feeding to the classifier. We center crop the image to 224 × 224 and normalize it before passing it to the classification network during inference.

Model architecture

We use ResNet-50 [30] pretrained on ImageNet [69] as the online encoder, and the target encoder, . The projector networks, are multi-layer perceptrons with a hidden layer consisting of 4096 neurons followed by batch normalization, ReLU activation, and an output layer of dimension 256. The predictor network, is architecturally the same as the projector. In our second stage of training, we modify the encoder block architecture to accommodate attention computation and initialize it with the pretrained weights from self-supervised training. We attach three attention estimators after ‘layer2’, ‘layer3’, and ‘layer4’ (layer names as per PyTorch implementation). The local features extracted at these three layers have dimensions (512, 28, 28), (1024, 14, 14) and (2048, 7, 7) respectively using ‘channel first’ representation. These three local features together with the global feature at the ‘avgpool’ layer produce three attended encodings. However, the global feature has a shape of (2048, 1), which causes shape incompatibility. To alleviate this issue, we use projector blocks consisting of 1 × 1 2-D convolution operations, which ensures that the channel dimension of local features matches the channel dimension of the global feature. Next, attention maps are computed using a linear combination of a local feature and the global features, 1 × 1 2-D convolution operations and softmax normalization. Finally, these attended embeddings are concatenated and classified using a linear classifier.

Hyperparameters

For self-supervised training, we use a batch size of , Adam optimizer with a learning rate of , and the model is trained for epochs. To train the classifier, we use a batch size of , Adam optimizer with an initial learning rate of with a cosine decay learning rate scheduler. Further, we use a global weight decay parameter of .

Computation complexity

For the self-supervised training we use 2 NVIDIA V100 GPU cards with 32 GB memory and 5120 CUDA cores in parallel. One epoch approximately takes 1.5 h to complete execution. For the finetuning stage, we use 1 NVIDIA V100 (32 GB 5120 CUDA cores) GPU card and one epoch takes approximately 8 min to complete execution. Confusion Matrix: The horizontal axis and the vertical axis correspond to the ground truth labels and the predicted classes respectively.

Quantitative results

To benchmark the proposed method against other state-of-the-art methodologies, we compute and report class-wise Precision (Positive Prediction Value), Recall (Sensitivity), F1 score, Specificity, Negative Prediction Value (NPV), and overall accuracy along with 95% confidence interval. Table 2 presents our findings. As can be seen from Table 2, the proposed method achieves the best overall accuracy with best 95% confidence interval. Further the proposed method achieves the best precision for COVID-19 cases meaning the proposed classifier rarely label a COVID-19 negative sample as a positive sample. Moreover, the proposed method achieves the best recall score implying the classifier is able to find most of the positive samples belonging to the COVID-19 class. The highest F1 score achieved by the proposed method indicates that the proposed method is the most balanced in terms of both precision and recall as compared to the baseline methods. Similarly, the proposed method achieves high specificity and NPV indicating that the false positive rate is low as well. Finally, from Fig. 3, it can be seen that the proposed method achieves best class-wise accuracy.
Table 2

Comparison of performance of the proposed method on chest X-ray dataset against state-of-the-art methods.

MethodClass labelPrecisionRecallF1-ScoreSpecificityNPVOverall accuracy (95% CI)
CoroNet [22]Normal0.91060.91450.91260.96440.96600.8932 (0.8701, 0.9135)
Pneumonia Bacterial0.86060.89260.87630.93990.9546
Pneumonia Viral0.92200.87840.89970.98370.9736
COVID-190.89340.88000.88660.96630.9617

COVIDNet [37]Normal0.91560.92740.92140.96610.97100.9078 (0.8859, 0.9266)
Pneumonia Bacterial0.88400.91320.89840.95020.9634
Pneumonia Viral0.93620.89190.91350.98670.9766
COVID-190.90820.89000.89900.97120.9650

Teacher Student Attention [39]Normal0.92740.91340.92030.97120.97110.9138 (0.8926, 0.9321)
Pneumonia Bacterial0.88890.92560.90690.95190.9685
Pneumonia Viral0.93710.90540.92100.98670.9794
COVID-190.91280.89000.90130.97280.9650

MAG-SD [38]Normal0.93990.93590.93790.97630.97460.9235 (0.9032, 0.9408)
Pneumonia Bacterial0.90360.92980.91650.95880.9704
Pneumonia Viral0.93750.91220.92470.98670.9809
COVID-190.91920.91000.91460.97440.9712

Proposed MethodNormal0.98670.95300.96960.99490.98160.9587 (0.9428, 0.9713)
Pneumonia Bacterial0.96170.93390.94760.98450.9728
Pneumonia Viral0.92160.95270.93690.98220.9896
COVID-190.95241.00000.97560.98401.0000
Fig. 3

Confusion Matrix: The horizontal axis and the vertical axis correspond to the ground truth labels and the predicted classes respectively.

Ablation studies to understand the impact of each training component. Visualization of different cases (Bacterial Pneumonia, Viral Pneumonia, and COVID-19) considered in this study and their associated critical factors in decision making by our proposed method. In each subfigure, the left figure presents the input to the model and its ground truth label; the right figure presents the predicted probabilities for each class and highlight the factors critical corresponding to the top predicted class. We have used jet colormap to colorize heatmap.

Ablation studies

In this section, we examine the impact of different training components proposed in this work. Specifically, we study the effect of pretraining on ImageNet [69], self-supervised pretraining on CheXpert [16] and the attention map. Table 3 presents the findings. When a ResNet-50 architecture is trained on the COVID-19 CXR dataset from scratch, its performance is the worst. Transfer learning (ResNet-50 [30] pretrained using ImageNet [69]) improves the model’s classification performance. Attention mechanism provides a further boost in performance. Finally, self-supervised pretraining using ChexPert [16] helps the model extract useful CXR specific features and enhance the model’s classification accuracy.
Table 3

Ablation studies to understand the impact of each training component.

Training components
Class labelPerformance metrics
Pretrained on ImageNet [69]Self-supervised learning on CheXpert [16]AttentionPrecisionRecallF1SpecificityNPVOverall accuracy
NoNoNoNormal0.86280.83330.84780.94750.93480.8483
Pneumonia Bacterial0.87720.82640.85110.95190.9295
Pneumonia Viral0.82170.87160.84590.95860.9715
COVID-190.82160.8750.84740.93910.9591

YesNoNoNormal0.88110.85470.86760.95420.9430.8786
Pneumonia Bacterial0.8970.86360.880.95880.9442
Pneumonia Viral0.85710.89190.87420.96750.9761
COVID-190.87140.9150.89270.95670.9723

YesNoYesNormal0.89910.87610.88740.9610.95130.8956
Pneumonia Bacterial0.90560.87190.88840.96220.9475
Pneumonia Viral0.86710.92560.89540.96890.9835
COVID-190.90240.9250.91360.96790.9758

NoYesNoNormal0.9080.92740.91750.96270.97090.915
Pneumonia Bacterial0.91770.92150.91960.96560.9673
Pneumonia Viral0.92410.90540.91470.98370.9794
COVID-190.91370.90.90680.97280.9681

NoYesYesNormal0.91630.93590.9260.96610.97440.9345
Pneumonia Bacterial0.95740.92980.94340.98280.9711
Pneumonia Viral0.93880.93240.93560.98670.9852
COVID-190.92610.940.93300.9760.9807

YesYesNoNormal0.92120.94870.93470.96780.97940.9454
Pneumonia Bacterial0.96640.95040.95830.98630.9795
Pneumonia Viral0.94560.93920.94240.98820.9867
COVID-190.94950.94000.94470.98400.9808

YesYesYesNormal0.98670.95300.96960.99490.98160.9587
Pneumonia Bacterial0.96170.93390.94760.98450.9728
Pneumonia Viral0.92160.95270.93690.98220.9896
COVID-190.95241.00000.97560.98401.0000

Qualitative results

Fig. 4 presents the attention map instrumental for the prognosis made by the proposed method. We present three visualizations one for each of bacterial pneumonia (Fig. 4(a)), viral pneumonia (Fig. 4(b)) and COVID-19 (Fig. 4(c)).
Fig. 4

Visualization of different cases (Bacterial Pneumonia, Viral Pneumonia, and COVID-19) considered in this study and their associated critical factors in decision making by our proposed method. In each subfigure, the left figure presents the input to the model and its ground truth label; the right figure presents the predicted probabilities for each class and highlight the factors critical corresponding to the top predicted class. We have used jet colormap to colorize heatmap.

Discussion and conclusion

This work introduces a method for automated COVID-19 prognosis using a limited amount of labeled COVID-19 CXR data. We have empirically demonstrated the effectiveness of the proposed method over existing SOTA methods as measured using various metrics such as precision, recall, F1 score, specificity, and NPV. While the proposed methodology is highly performant, it is not error-free as the CXR findings due to COVID-19 are not exclusive and overlap with other thoracic infections [70]. Therefore, to improve the efficiency of diagnosis and efficient resource utilization, we suggest the proposed method to be used in conjunction with RT-PCR, and first-line treatment may be initiated based on CXR findings while the RT-PCR test report is awaited. Despite, the great success achieved by deep learning models in different machine learning tasks, they are prone to various biases such as selection bias (distribution of training examples is not reflective of their real-world distribution), group attribution bias (tendency to generalize what is true of individuals to an entire group to which they belong) and so on. Therefor, to deploy the proposed method clinically, it is imperative to thoroughly evaluate the model through clinical trials to examine its generalization capabilities and stability. Although, the method proposed in this work is highly performant on multinational dataset (since the datasets used in this study were compiled from several repositories), to further improve the generalization ability of the proposed method, the model needs to be trained on a large, diverse, high-quality dataset. To conclude, preventing the spread of COVID-19 requires early diagnosis. While RT-PCR is highly accurate when the test is conducted appropriately, its turn-around time is high. Therefore, our proposed deep neural framework might be useful to initiate the first line treatment. Further, the proposed method, when used in conjunction with RT-PCR can be thought of as a complimentary diagnosis or a second opinion to ensure efficient utilization of limited resources. In our future work, we intend to extend this work to automate the analysis of infection severity.

CRediT authorship contribution statement

Arnab Kumar Mondal: Conceptualization, Data curation, Methodology, Formal analysis, Validation, Software, Writing – original draft, Writing – review & editing.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  34 in total

1.  Thorax-Net: An Attention Regularized Deep Neural Network for Classification of Thoracic Diseases on Chest Radiography.

Authors:  Hongyu Wang; Haozhe Jia; Le Lu; Yong Xia
Journal:  IEEE J Biomed Health Inform       Date:  2019-07-12       Impact factor: 5.772

2.  Triple attention learning for classification of 14 thoracic diseases using chest radiography.

Authors:  Hongyu Wang; Shanshan Wang; Zibo Qin; Yanning Zhang; Ruijiang Li; Yong Xia
Journal:  Med Image Anal       Date:  2020-10-16       Impact factor: 8.545

3.  Coronavirus covid-19 has killed more people than SARS and MERS combined, despite lower case fatality rate.

Authors:  Elisabeth Mahase
Journal:  BMJ       Date:  2020-02-18

4.  COVID-19 Automatic Diagnosis With Radiographic Imaging: Explainable Attention Transfer Deep Neural Networks.

Authors:  Wenqi Shi; Li Tong; Yuanda Zhu; May D Wang
Journal:  IEEE J Biomed Health Inform       Date:  2021-07-27       Impact factor: 7.021

5.  Multiscale Attention Guided Network for COVID-19 Diagnosis Using Chest X-Ray Images.

Authors:  Jingxiong Li; Yaqi Wang; Shuai Wang; Jun Wang; Jun Liu; Qun Jin; Lingling Sun
Journal:  IEEE J Biomed Health Inform       Date:  2021-05-11       Impact factor: 7.021

6.  New machine learning method for image-based diagnosis of COVID-19.

Authors:  Mohamed Abd Elaziz; Khalid M Hosny; Ahmad Salah; Mohamed M Darwish; Songfeng Lu; Ahmed T Sahlol
Journal:  PLoS One       Date:  2020-06-26       Impact factor: 3.240

7.  COVID-Net CXR-2: An Enhanced Deep Convolutional Neural Network Design for Detection of COVID-19 Cases From Chest X-ray Images.

Authors:  Maya Pavlova; Naomi Terhljan; Audrey G Chung; Andy Zhao; Siddharth Surana; Hossein Aboutalebi; Hayden Gunraj; Ali Sabri; Amer Alaref; Alexander Wong
Journal:  Front Med (Lausanne)       Date:  2022-06-10

8.  A deep learning approach to detect Covid-19 coronavirus with X-Ray images.

Authors:  Govardhan Jain; Deepti Mittal; Daksh Thakur; Madhup K Mittal
Journal:  Biocybern Biomed Eng       Date:  2020-09-07       Impact factor: 4.314

9.  Coronavirus disease (COVID-19) detection using X-ray images and enhanced DenseNet.

Authors:  Saleh Albahli; Nasir Ayub; Muhammad Shiraz
Journal:  Appl Soft Comput       Date:  2021-06-25       Impact factor: 6.725

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.