Literature DB >> 35698723

Covid-MANet: Multi-task attention network for explainable diagnosis and severity assessment of COVID-19 from CXR images.

Ajay Sharma1, Pramod Kumar Mishra1.   

Abstract

The devastating outbreak of Coronavirus Disease (COVID-19) cases in early 2020 led the world to face health crises. Subsequently, the exponential reproduction rate of COVID-19 disease can only be reduced by early diagnosis of COVID-19 infection cases correctly. The initial research findings reported that radiological examinations using CT and CXR modality have successfully reduced false negatives by RT-PCR test. This research study aims to develop an explainable diagnosis system for the detection and infection region quantification of COVID-19 disease. The existing research studies successfully explored deep learning approaches with higher performance measures but lacked generalization and interpretability for COVID-19 diagnosis. In this study, we address these issues by the Covid-MANet network, an automated end-to-end multi-task attention network that works for 5 classes in three stages for COVID-19 infection screening. The first stage of the Covid-MANet network localizes attention of the model to the relevant lungs region for disease recognition. The second stage of the Covid-MANet network differentiates COVID-19 cases from bacterial pneumonia, viral pneumonia, normal and tuberculosis cases, respectively. To improve the interpretation and explainability, three experiments have been conducted in exploration of the most coherent and appropriate classification approach. Moreover, the multi-scale attention model MA-DenseNet201 proposed for the classification of COVID-19 cases. The final stage of the Covid-MANet network quantifies the proportion of infection and severity of COVID-19 in the lungs. The COVID-19 cases are graded into more specific severity levels such as mild, moderate, severe, and critical as per the score assigned by the RALE scoring system. The MA-DenseNet201 classification model outperforms eight state-of-the-art CNN models, in terms of sensitivity and interpretation with lung localization network. The COVID-19 infection segmentation by UNet with DenseNet121 encoder achieves dice score of 86.15% outperforming UNet, UNet++, AttentionUNet, R2UNet, with VGG16, ResNet50 and DenseNet201 encoder. The proposed network not only classifies images based on the predicted label but also highlights the infection by segmentation/localization of model-focused regions to support explainable decisions. MA-DenseNet201 model with a segmentation-based cropping approach achieves maximum interpretation of 96% with COVID-19 sensitivity of 97.75%. Finally, based on class-varied sensitivity analysis Covid-MANet ensemble network of MA-DenseNet201, ResNet50 and MobileNet achieve 95.05% accuracy and 98.75% COVID-19 sensitivity. The proposed model is externally validated on an unseen dataset, yields 98.17% COVID-19 sensitivity.
© 2022 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  Chest X-ray; Covid-19; Deep learning; Explainable AI; Infection segmentation; Lung segmentation; Transfer learning

Year:  2022        PMID: 35698723      PMCID: PMC9170279          DOI: 10.1016/j.patcog.2022.108826

Source DB:  PubMed          Journal:  Pattern Recognit        ISSN: 0031-3203            Impact factor:   8.518


Introduction

Coronavirus disease 2019 (COVID-19), has been declared a global epidemic by WHO in early March 2020. Globally, as of 11 August 2021, a corpus of 204,644,849 confirmed cases including 4323,139 deaths have been reported to WHO because of COVID-19 infection caused by SARS CoV-2 [1]. Still, these numbers are increasing and predicted to grow rapidly [12] in the upcoming months because of the higher reproduction rate of the disease. Therefore, it is essential to lessen the spread of virus which is possible only by early detection, treatment, and isolation of virus cases. In general, three major screening methods used for early diagnosis of COVID-19 include, reverse transcriptase-polymerase chain reaction (RT-PCR), Chest X-ray (CXR) and Chest computed tomography (CT) [2]. The standard COVID-19 diagnosis RT-PCR method detects virus RNA from nasopharyngeal swab or sputum but it requires expert personnel, specific material, and laboratories for testing and is a time-consuming process. Another alternative to the RT-PCR test is the rapid antigen test, which gives a faster and less expensive diagnosis compared to PCR but has poor sensitivity to COVID-19 [3]. In particular, radiological examinations were found very useful in the diagnosis and assessment of COVID-19 disease progression. Some recent research studies were conducted by expert radiologists on sensitivity analysis of COVID-19 by RT-PCR, CT, and CXR scans. Since preliminary 3-days clinical practice performed on CT and RT-PCR test involving 51 patients describes with CT images a sensitivity of 98% outperforming RT-PCR having 71% sensitivity [4]. However, due to the rapid surge in COVID-19 cases, routine use of chest CT is not feasible because of its portability and expensive setup. Another study by wong et al. [12], involving 64 participants, reported lower sensitivity of 69% using the CXR modality compared to RT-PCR having 91% initial sensitivity. Despite having low sensitivity, more than 9% of negatively reported RT-PCR tests have shown certain abnormalities of COVID-19 in CXR radiographs. Although the evaluation by CXR is less accurate in the early studies compared to CT and RT-PCR. Still, it is considered an efficient and standard screening tool because of its low-cost, minimally-invasive, quick results, and requires simpler logistics for its implementation [16]. In addition to COVID-19, other common lung diseases like viral pneumonia, bacterial pneumonia, and tuberculosis possess almost similar symptoms as observed in COVID-19 pneumonia. Thus, it is essential to develop a system that uniquely identifies COVID-19 patients among these common lung diseases. Fig. 1 displays example samples for each disease class considered in this study.
Fig. 1

Example sample images; (a) COVID-19 (b) Normal (c) BP (d) VP (e) TB classes.

Example sample images; (a) COVID-19 (b) Normal (c) BP (d) VP (e) TB classes. Specifically, in this decade (2011–2021) deep learning approaches have been successfully explored for image classification and segmentation tasks. These approaches play a vital role in analyzing CXR radiographs as a standard tool for the early classification of COVID-19 disease. In 2012, the first advanced deep learning model AlexNet [5] was proposed that introduces overlap pooling, ReLU, dropout, large 11×11, and 5 × 5 kernels. In the meantime, evolution in deep learning approaches rapidly increases after AlexNet won the ImageNet [10] challenge with a 16.4% error rate. Some advanced deep learning architectures proposed include VGG [6] in 2014 uses a small 3 × 3 kernel, GoogleNet [6] in 2015 introduced the block concept, split and merge idea with different kernel sizes. Inception [7] in 2015 modified the GoogleNet model and introduced the split, transform and merge concept with asymmetric filters. ResNet [7] in 2016 introduced residual learning with skip connections, MobileNet [6] in 2017 introduced depth-wise convolution [8] followed by pointwise convolution [9], and DenseNet [11] in 2017 introduced cross-layer connections for information flow. The evolution of these deep learning models [12] improve the classification ability continuously when tested on benchmark datasets [13]. Some application areas [14] where these CNN models have been successfully employed for early diagnosis include cancer [15] classification [16], skin lesions classification [17], and the diabetic retinopathy diagnosis [18], etc.

Literature review

Recently, researchers proposed numerous AI-based methods for the detection of COVID-19 using chest x-ray and CT images. Even though the usage of deep learning approaches gave promising results having higher statistical scores. Still, these works need desirable improvements at the model development stage to achieve an explainable diagnosis for classification and infection region segmentation. For instance, the first research work proposed by Wang et al. [19] indicates that the application of deep learning models successfully identifies COVID-19 signs in CXR scans. Models successfully predict 80% correct results for the COVID-19 class. However, the reduction of false negatives is essential in building a reliable diagnosis system. In research work [20], the author projected COVID-SDNet methodology that tends to improve false negatives and classify COVID-19 into different severity labels such as mild, moderate, severe, normal-PCR+. Models composed COVIDGR-1.0 dataset available publicly and introduced segmentation-based cropping by discarding areas outside lung regions without lung segmentation in the raw image. In addition, transformation and augmentations of CXR scans based on the GAN model improved ResNet50 performance with an accuracy of 97.72%, 86.90% in severe, moderate cases. COVID-DeepNet, [21] hybrid multimodal model proposed for COVID-19 diagnosis using CXR images. Image-preprocessing with CLAHE and Butterworth bandpass filter eliminates noise and enhances contrast. Results of CNN and deep belief networks are fused for final class prediction among two classes with COVID-19 sensitivity of 99%. In the study [22], Shamsi proposed a transfer learning-based deep uncertainty aware model for CT and CXR analysis. The proposed study extracts feature by four DL approaches, InceptionResNetV2, ResNet50, VGG16, and DenseNet121, each passed to eight machine learning models for recognition of COVID-19. The statistical analysis reported SVM and neural net have higher accuracy and sensitivity ranging from 88% to 97% for the two-class problem. Research study by Wang et al. [23] proposed a classification model involving two stages of classification, named as Discrimination-DL stage extracting lung features and Localization-DL stage to localize features in left, right lungs regions. This network reported 98.71% accuracy of Discrimination-DL and 93.01% for localization network. In a study [24], the author developed a Graph-based approach for COVID-19 disease identification with minimal supervision required. This semi-supervised learning framework deals with vast unlabelled samples by pseudo-labeling. Results outperformed the supervised framework with better identification based on attention maps. In the reference study [25], the ConvNet model is developed based on fuzzy logic and deep learning models. Model improved learning when features extracted by a hybrid of deep and fuzzy logic approach given to multilayer perceptron for classification achieve 81% accuracy. Further developments in the state-of-the-art involve lung segmentation, image enhancement, and the creation of localization maps by Grad-CAM to make models explainable. In a research study [26], the author explored the effect of image enhancement on COVID-19 diagnosis using CXR data. Various techniques like CLAHE, histogram equalization (HE), gamma correction, image complement, and BCET were tested on raw and segmented images. It created a dataset of 18,479 images with classes as normal, COVID-19, and pneumonia. The experimental study uses InceptionV3, ChexNet, ResNet50, ResNet101, and DenseNet201 as CNN models, where ChexNet and DenseNet201 achieve 96.29% accuracy when applied with gamma correction. Another model developed by author [27] CMTNet for recognition of COVID-19 patients and finding regions infected by a disease. The proposed model follows VGG19 backbone network for the encoder model. Two separate decoder branches for lung and disease segmentation shares a common encoder branch. The study compiled an annotated dataset of 9000 CXR for lung segmentation and 200 for disease localization. Statistical analysis gains 87.20% sensitivity and 96.80% precision. In [28], the author improved false negatives and interpretation of the COVID-Net model by tuning existing parameters. The study recognizes 87% of COVID-19 samples correctly among pneumonia and normal classes. This modified model is tested on a large corpus of 8573 COVID-19 samples. The statistical analysis shows improvement in accuracy and interpretation after lung segmentation to 91% with 87% sensitivity. In such epidemic situation, data collection is the biggest challenge to train deep learning models. In research work [29], the author proposed a transfer learning, ResNet model with modified patch-wise training and lung segmentation approach to mitigate limited dataset challenge. The model was trained to classify 5 classes i.e., COVID-19, normal, viral pneumonia, bacterial pneumonia, and TB. Lung contours are segmented with the FC-DenseNet model and ResNet-18 to classify segmented lungs among one of the pathology classes with 89% accuracy and 85% sensitivity, respectively. Ghoshal and Tucker, [30] explored the uncertainty of deep learning models based on Spearman correlation and bayesian uncertainty. Usage of bayesian CNN with drop weights and ResNet50v2 improved diagnostic interest and explainability of class activation maps. Shi et al. [31] developed EXAM, an explainable attention-based model generating results explaining the diagnostic interest of models for classification by Grad-CAM for 3-classes. EXAM developed attention by merging spatial and channel-wise features with DenseNet models. Kumar and Singh et al., [32] proposed a deep learning model for distinguishing three classes COVID-19, normal, and pneumonia. Image segmentation, and image enhancement techniques are used to develop a stacked ensemble of four CNN models using Naïve Bayes as a meta learner and Grad-CAM is used for qualitative interpretation investigation. In the study [44], Mangal et al., proposed an explainable CovidAID model diagnosing COVID-19 from viral pneumonia, bacterial pneumonia, and normal cases using the backbone model DenseNet121 outperforming COVID-Net with an accuracy of 90.5% and sensitivity of 100% with 182 PA view COVID-19 samples. Another research work done by Wang et al. [33] makes a standardized DL pipeline for the classification of COVID-19 pneumonia and lesion visualization for diagnosis. The standardization stage discards irrelevant features outside lung regions by capturing lung regions. The statistical analysis measures AUC score of the system ranges from 0.87 to 0.97 for COVID and other or viral pneumonia,0.87 for non-severe and severe COVID-19, between 0.94–0.98 for viral and other pneumonia types. In the study [34], Tahir et al. proposed a model for infection quantification and grading of COVID-19 pneumonia. The study compiled a large dataset of 33,920 samples having 11,956 COVID-19 samples in original and augmented form. Statistical analysis compared results of segmented, non-segmented lungs and infection by state-of-the-art UNet, UNet++, and FPN (Feature pyramid learning) models. These models utilize ResNet, VGG19, DenseNet121, etc. as the backbone of encoder-decoder networks. Model localized COVID-19 infection regions with an overall Dice score of 88%. Signoroni et al. [35] proposed BS-Net model providing a deep learning-based framework for the diagnosis of COVID-19 in CXR. Model localize infection regions by assigning quantitative scoring to lung regions based on the Brixia scoring system. Weakly supervised learning is used to achieve tasks such as segmentation, score-estimation, and spatial alignment. BS-Net tested on 5000 CXR segmented images based on the UNet variant model with a 94% IoU score. Research work proposed by Gidde et al. [36] developed CovBaseAI explainable decision system by an ensemble of three deep learning models for COVID-19 diagnosis. Model validation is performed by 2 datasets having a corpus of 471 and 1401 for COVID-19/Normal CXR scans. The statistical analysis achieved an accuracy of 87% accuracy with 98% negative predictive value. In addition to standard CXR modality, some research models have been developed using CT, ultrasound modality, or using multimodal data involving CXR and CT images both. But in real practice testing using CT images lacks portability and is expensive compared to the CXR modality. In this context, Owais et al. [37] proposed a lightweight deep learning ensemble network by combining FCNet, ShuffleNet, and MobileNet for COVID-19 diagnosis. Localization and activation map visualization enables radiologists to diagnose infection focused by model. However, the model is tested with a collection of CXR and CT datasets with a mean F1-score of 95.94% and 94.60%. The area under curve achieved is 97.99% and 97.50%, respectively. Another model [38] proposed for diagnosing community-acquired pneumonia from 3D CT volumes using a dual-sampling attention network. Dual sampling resolves class imbalance issue and attention mechanism better identifies infection in CT volumes with 3D CNN model outperforming state-of-the-art UNet model. The study processed a corpus of 2186 CT scans with labels COVID-19/NORMAL by 5-fold cross-validation. Statistical analysis attained AUC, F1-score, and accuracy of 0.944, 82%, and 87.5%, respectively. The majority of research studies involve CXR and CT modalities for the diagnosis of COVID-19 by application of transfer learning models. However, HORRY and CHAKRABORTY [39] research work developed a multimodal model for diagnosis among multiple modalities by similar transfer learning approaches. In addition to CT and CXR scans, ultrasound modality has also proven useful in the diagnosis process but has portability and cost issues for large masses. Study optimized VGG19 model, compared with other popular CNN models DenseNet, Inception, VGG16, and ResNe50 for statistical analysis. Statistical measures reveal Ultrasound modality gives superior results when compared with CT and CXR images for the 3-class classification problem. Precision values achieved are 86% for X-ray, 84% for CT, and 100% for ultrasound values having F1-score of 99%. However, the model is a black box lacking explanations and localization visualization maps using whole slide images. Another explainable research model [40] JCS proposed for the diagnosis of COVID-19 in chest CT images. JCS framework involves joint segmentation and classification to identify COVID-19 infection in CT images. The proposed work compiled a large dataset of 144,167 CT images from 350 normal CT scans and 400 COVID-19 patients. For infection quantification, 3855 CT images are annotated named COVID-CS dataset. Joint diagnosis achieves sensitivity and specificity of 95% and 93%, respectively, for classification whereas 78.5% dice score for infection segmentation.

Issues affecting results in the literature

In this section, we discuss certain shortcomings after critically analyzing state-of-the-art developments associated with existing diagnosis approaches. Despite having good results of existing methodologies still, these approaches lack explainable diagnosis since the interpretation of the model's decision is not correlated to lung regions. The major challenge with existing methodologies lies in weaker interpretation and low explainability of models for the classification of disease. Some state-of-the-art models created only backbox, not explaining whether the model focused on relevant lung regions for classification. In the beginning, models applied directly on whole slide CXR radiographs gave poor interpretation enabling classification based on out-of-lungs region. Some existing studies performed segmentation of lungs based on the UNet model, shows improvement in interpretation but there is a minor accuracy drop in these approaches. However, most of the works show a good correlation to disease but signify weaker diagnostic interest to lung regions. Indeed, the proposed segmentation-based cropping approach represents the strong correlation of disease classification to the lung regions. Secondly, most studies revolve around the classification of two or three classes distinguishing COVID-19 from normal and pneumonia classes [19]. However, tuberculosis and virus pneumonia have similar symptoms as seen in COVID-19 but after extensive searches, only two studies were found involving five classes for classification. These studies lack interpretation and explainable nature [22] of models working with limited example datasets. However, in clinical practice, explainable diagnostic system essentially have infection region quantification and severity assessment capability in addition to detection that is missing. Additional experimental issues in the state-of-the-art as per recommendations of work published in nature [41] have been considered to avoid pitfalls in classification tasks. The proposed study strictly considers both mandatory and non-mandatory recommendations for deep learning modeling to pass the quality screening procedure. The major issues responsible for the failure of quality screening procedure are; no explanations provided for final model selection, lack of image pre-processing steps, no clear specifications for training approach, class imbalance, absence of robustness and sensitivity analysis, no external validation, lack of performance analysis on confidence threshold intervals, interpretation and generalization. In brief, some published works present unclear implementation details such as; initial preprocessing steps [23] for conversion of the dataset into a uniform format and the range is either missing or not specified. In addition, specifications of training parameters like an optimizer, learning rate, loss function [31], and reason for selection of the best model are missing [54]. In the literature survey, all the studies involve two or more classes but few performed class-wise sensitivity analysis [35] for better behavior analysis of the model [36] for each of the classes. Despite class imbalance majority of studies uses categorical class entropy loss [36] that may prioritize the learning of the majority class [55] only. In the proposed research work, we consider all these challenges to develop a final multi-task model improving interpretation and generalization by segmentation-based cropping, explainable diagnosis by infection quantification model, localization map by Grad-CAM, and class imbalance by the multilabel classification loss function. The model robustness is maintained by a clear experimental setup, pre-processing steps, class-wise sensitivity analysis, confidence analysis, and external validation.

Research contributions

The early success of deep learning methods for COVID-19 diagnosis motivates us to further investigate CNN to minimize existing issues and improve the interpretation and explainability of diagnosis. We propose the Covid-MANet model that enables diagnosis and progression of the disease by analyzing relevant features in lung regions extracted using segmentation masks. Covid-MANet is a three-stage deep learning model that benefits from lung localization, segmentation-based classification, data-augmentation, pre-processing, post-processing, and multi-label loss to curtail any variances and imbalance among CXR radiographs collected from different repositories. Covid-MANet works for the 5-way classification problem, differentiating COVID-19 from bacterial pneumonia (BP), viral pneumonia (VP), normal, and tuberculosis (TB) cases. Precisely, the first stage in Covid-MANet takes Raw CXR radiographs as inputs, uses a segmentation model, makes a prediction of CXR lung masks and performs the post-processing of predicted masks for localization of the lung region. This forces model to attention more on segmented regions resulting in improved interpretation and generalization. In addition, segmentation models are trained with new pseudo RANZCR dataset in combination with existing datasets to enable better segmentation of VP and BP cases. In the second stage, the MA-DenseNet201 detection model improves its generalization by concentrating more on relevant lung regions extracted by the first stage. Segmentation-based cropping is conducted to further improve the interpretation of models. The final stage of the Covid-MANet model performs quantification and severity assessment of COVID-19 infection in the lungs that assist doctors to understand the progression of the disease. The severity levels assigned to the disease are mild, moderate, severe, and critical as per the RALE scoring system. In the end, Grad-CAM is used to interpret the features focused by the model for the classification of an image to a particular class in each experiment. To accomplish segmentation, UNet, UNet with residual block, and UNet with Dense block are used but UNet with dense block achieves better segmentation of lung regions compared to the other two. To perform classification in these experiments, the proposed model is compared with state-of-the-art CNN models, VGG16, VGG19, InceptionV3, MobileNet, ResNet50, NASNetMobile, DenseNet121, DenseNet201 based on performance in the existing methodologies and ImageNet database. The proposed infection segmentation system uses UNet with DenseNet121 encoder for quantification of infection region in COVID-19 classified samples. The major contributions of this study are summarized as follows: The Covid-MANet is a single generic multi-task framework for automated COVID-19 diagnosis, infection region quantification and severity assessment of COVID-19 into more specific levels as mild, moderate, severe, and critical. The proposed work provides an enhanced segmentation-based classification model with dense blocks where modified UNet architecture is investigated with two other SOTA models for automated segmentation of lung regions in CXR images. The large dataset is created by the inclusion of the RANZCR dataset that provides better segmentation of unseen VP and BP images. The Covid-MANet model improves generalization and interpretability for COVID-19 classification by introducing segmentation-based cropping and classification by the MA-DenseNet201 model with multiscale attention network outperforming state-of-the-art networks. In addition, this end-to-end framework not only classifies but also segment infection region aimed at screening the progression of the disease. The proposed work investigates the class-wise sensitivity analysis in three experiments. Based on prior awareness of various class level accuracies, we propose a weighted average ensemble approach (WAE) that outperforms state-of-the-art models for all the classes. Finally, a gradient-weighted class activation mapping (Grad-CAM) is used for explainable diagnosis to generate a localization map for each disease type investigates model interpretation in addition to COVID-19 infection map. The segmentation-based cropping approach reduces all biases and develops a generalizable model more stable as compared to whole slide and segmented images. The structure of the whole research study is organized as; Section 2 discusses the detailed framework of the proposed model involving lung segmentation, classification and infection segmentation. Section 3 discusses the dataset resources utilized, experimental setup, and evaluation metrics for segmentation and classification. Section 4 discusses the quantitative and qualitative analysis of the proposed methodology and Section 5 ends with a conclusion.

Covid-MANet multi-task framework

This section presents the methodology of the Covid-MANet network developed for the diagnosis and infection quantification of COVID-19 in CXR samples. The primary focus of the multi-task network is to develop an explainable diagnosis system representing a higher correlation of disease classification to the lungs region. In the first task, the segmentation network localizes lung areas ready for input to the classification model based on segmentation-based cropping. Then, the proposed classification model classifies the image into one of pathology class and the infection segmentation network quantifies the severity of COVID-19 infection into mild, moderate, severe, or critical. The detailed discussion of the multi-task network and loss functions for segmentation and classification is explained in the following sections:

Lung segmentation network and data processing

The first task of the Covid-MANet network is to segment out lung regions since the relevant disease information lies only in the lung regions. Recall that the proposed model works for five classes, differentiates COVID-19 from other lung diseases. However, no public dataset available in all five classes in one dataset. So, a compilation of the dataset by the fusion of the COVID-19 class with other lung disease resources may possess different acquisition conditions and artifacts that affect classification results. Classification based on lung segmentation resolves all these biases, therefore resulting in higher interpretation and generalization of the model. The automated lung segmentation task achieved by UNet and its variants inspired by ResNet and DenseNet model. Fig. 2 shows the generic architecture of the segmentation network and its blocks. Fig. 2(a) represents a general encoder-decoder model for segmentation where skip connections connect encoded features to the decoder. Three segmentation models can be constructed by replacing the 3 × 3 block of a general model with structural blocks (b–d). Fig. 2(b) displays 2 convolutional layers of the UNet block [42], Fig. 2(c) represents a basic block of ResUNet model [43], Fig. 2(d) displays the building block [43] of DenseUNet model. The training of segmentation models is achieved by hybrid segmentation loss function [45] and the best model selection by IoU and Dice coefficient metrics. The references for compilation of a larger dataset compared to other models specified in Section 3.1, containing images and labeled lung annotated masks for supervision. However, the lung masks for COVID-19, TB, and normal classes are publicly available, but VP and BP classes have no annotated masks. So, another publicly available RANSCR dataset is utilized that has a similar UID compared to VP and BP, enabling better lung localization of unsupervised BP and VP classes.
Fig. 2

Architecture of lung segmentation network, (a) basic UNet type model where 3 × 3 block is replaced by block b, c or d (b) Convolutional block of UNet (c) Residual convolutional block of ResUNet (d) Dense convolutional block of DenseUNet model.

Architecture of lung segmentation network, (a) basic UNet type model where 3 × 3 block is replaced by block b, c or d (b) Convolutional block of UNet (c) Residual convolutional block of ResUNet (d) Dense convolutional block of DenseUNet model.

Pre-processing and post-processing

Pre-processing is an essential step to achieve uniformity of each input class required for modeling since the dataset compiled by a fusion of different classes that may represent non-uniformity in size, range, or datatype. Pre-processing operations applied before segmentation includes resizing of input to 320 by 320, change of datatype to float32, and range consistency by normalization to [0,1]. The augmentation operations are applied to increase the size of the dataset and avoid overfitting. These operations are, image shifting range [−5.25%, −5.25%] vertically and horizontally, image rotation by [−15, 15] angle, Zoom by [0.05, 0.05] range. Augmentation operations are considered while training the model whereas model evaluation is free from augmentation. The proposed DenseUNet model outperformed ResUNet and UNet model in lung localization task with higher dice and IoU score. DenseUNet model is considered best for lung localization of unseen classification task. In addition, post-processing operations are applied to eliminate minor defects in the prediction of unseen images inside or outside ROI to improve lung localization. Three post-processing operations applied in a sequence using the Scikit-image library. These operations include first labeling the predicted mask, discarding small objects inside or outside the lungs by keeping the two largest lung regions, and finally preserving edges by dilation operation with a 5 × 5 kernel. Fig. 3 shows lung localization maps for each class, where predicted lung masks are post-processed to eliminate artifacts out-of-lung region.
Fig. 3

Example of segmented CXR samples, (a) COVID-19, (b) Normal, (C) BP, (d) VP, (e) TB, where rows correspond to the original pre-processed image, predicted lung masks, post-processed lung masks and final segmented lung contour.

Example of segmented CXR samples, (a) COVID-19, (b) Normal, (C) BP, (d) VP, (e) TB, where rows correspond to the original pre-processed image, predicted lung masks, post-processed lung masks and final segmented lung contour.

Classification network and data processing

The second stage of the Covid-MANet network aims to classify the input CXR sample into one of five pathology classes based on the correct interpretation to the lung localization area. To improve interpretation and generalization three experiments have been conducted with and without lung segmentation. The classification network adopted the proposed multi-scale attention model named MA-DenseNet201 using DenseNet201 [11] as a backbone architecture that acts as a feature extractor. The reason for using a pre-trained backbone network is to avoid overfitting, better training and extraction of features even from small datasets. The proposed model is compared with state-of-the-art pre-trained deep learning networks such as VGG16, VGG19, MobileNet, ResNet50, InceptionV3, DenseNet121, DenseNet201, and NASNetMobile [6] in each of the experiment. Once lung segmentation masks of the classification dataset are ready to use after post-processing are masked with the corresponding classification image to obtain the region of lungs. These lung segmented images are given to the classification network with and without a segmentation-based cropping approach. The classification network aims to improve interpretability and generalization by training and evaluation based on three distinct experiments. The qualitative analysis measuring the interpretation of each class type based on the best model in each experiment is shown in Fig. 5.
Fig. 5

Visualization of Grad-CAM maps for each class, where the first row corresponds to input images. The second, third and fourth rows show Grad-CAM activation visualization produced by the best classification model in each of the experiment indicating proposed segmentation-based cropping approach has better interpretable results.

The detailed analysis of experiments and processing steps followed to improve interpretation is discussed as follows;

Experiment 1: raw data

In experiment 1, classification models are trained on whole slide images without the involvement of the lung localization model. The input to the classification model is pre-processed, where images are resized and normalized to a fixed range (0,1). The visualization map generated by Grad-CAM explains result of this experiment are least interpretable as the model focuses more on artifacts or information present outside lung regions.

Experiment 2: lung localization

The second experiment has been conducted to resolve the difficulty that arises in the first experiment by the addition of a lung localization model. The lung localization model eliminates artifact outside the lungs enabling better classification based on the interesting region of the lungs. To achieve this, classification data is first given to the stage 1 lung localization model generating post-processed lung masks. The post-processed masks are masked with corresponding raw images to zero out information outside lung regions, generating lung segmented images. These segmented images are given to classification models for training, after resizing and normalization to a consistent range. However, some models are trained with three different versions in search of the best suitable approach. Initially, lung localization images are directly given as input to the classification model. Secondly, the application of the attention module CBAM involves channel attention followed by spatial attention applied before classification layers in a residual manner. Finally, the image enhancement approach gamma correction is applied to segmented images before being given to the classification model. Experiment 2 improved the interpretation of models, but few images make incorrect interpretations considering the region outside the lungs.

Experiment 3: segmentation based cropping

The proposed experimental approach resolves the difficulties of experiment 2 with maximum interpretation among all three experiments. This approach uses post-processed lung localization maps generated by stage 1, masked with original images, and finds a convex hull to completely discard edges. Generation of best-fit convex hull delimits left, top, right, and bottom regions of segmented lungs. The convex hull is computed by first converting the segmented image to a grayscale image, applying thresholding operation to the grayscale image, and detecting extreme points by contour detection. Finally, extreme points are utilized to crop segmented lungs ready for classification. Cropped images are given to the classification model for disease identification just after normalization in the range (0,1). Similarly, this approach is trained on four different versions explained as; Initially, cropped images are directly given to classification models. Secondly, cropped images are enhanced by gamma correction before given to the classification model. In the third version, attention mechanism with CBAM is applied in addition to gamma correction before the classification layers. Finally, cropped images are denoised using a total variation filter followed by gamma correction to generate data for classification. After extensive experiments, we found proposed segmentation-based cropping approach gives maximum interpretation as shown by Grad-CAM activation maps. The Grad-CAM localization maps indicate better interpretation in experiment 3 corresponding to all the classes involved in the study.

Multiscale attention-based classification model

Attention and multiscale feature extraction has been used in several studies to converge model attention to more relevant regions and discard less important features. We developed an enhanced classification model considering attention maps of different scales, guiding the model in the feedforward process. Multiscale hybrid attention module with transfer learning backbone makes better decisions based on high-level feature maps. DenseNet201 model is considered as backbone because of its deep feature extraction ability for better COVID-19 diagnosis. The proposed hybrid attention generation model aims at extracting attention maps from multiple-scale feature maps. From the DenseNet201 pre-trained model, feature maps of shapes 14×14×1024, 7 × 7*1856, and 7 × 7*1920 are used to make a corpus of 32 attention maps. The schematic representation of the MA-DenseNet201 model and multiscale hybrid attention generator is shown in Fig. 4 . Multiscale hybrid attention generator module extracts feature maps f1, f2, f3 from backbone model, generates attention A1, A2, A3 by 1 × 1 convolution operation. All the feature maps are downsampled to 7 × 7 feature maps. These attention maps connect residually to generate hybrid attention map A of size 32. The feature maps generated by the DenseNet201 backbone network concatenate with a multiscale hybrid attention generator. Finally, the output map is processed by a 1 × 1 convolution layer. The classification layers added on top of this network include flatten layer, dense layer followed by a dropout layer, and final dense layer with 5 neurons and a softmax activation function.
Fig. 4

The architecture of MA-DenseNet201 model adding multiscale hybrid attention module to DenseNet201 backbone.

The architecture of MA-DenseNet201 model adding multiscale hybrid attention module to DenseNet201 backbone. Visualization of Grad-CAM maps for each class, where the first row corresponds to input images. The second, third and fourth rows show Grad-CAM activation visualization produced by the best classification model in each of the experiment indicating proposed segmentation-based cropping approach has better interpretable results.

COVID-19 infection quantification and severity assessment

In clinical diagnosis, end-to-end explainable model development not only requires blackbox classification. But it is desirable to have an automated infection segmentation module measuring the progression of COVID-19 disease [34]. However, COVID-19 infection segmentation in the lungs is much more complex compared to lung segmentation. For infection region segmentation, we adopted UNet based encoder-decoder model [42] with a dense backbone network. UNet model with DenseNet121 backbone acts as an infection segmentation model, which is compared with seven other models involving UNet, Attention-UNet, UNet++, R2UNet [45] by considering backbone architecture as VGG19 with UNet++, ResNet50, and DenseNet201 with UNet. Similar to classification, the infection segmentation model is tested on two scenarios as of classification. Furthermore, for a better understanding of the progression of infection in the lungs, COVID-19 infection is quantified based on the predicted lungs region. The quantification of infection in the lungs is computed by the sum of predicted disease pixels divided by the total number of pixels in the lungs. In addition, infection quantification in left and right lungs is computed separately. Finally, infection score assignment and severity assessment module are introduced based on RALE [35] scoring system. According to the RALE scoring system, lungs are divided into two separate regions left and right lungs. Each lung is assigned a score between 0 and 4 based on the percentage of infection found. The score 0 is assigned for no infection involvement, score 1 for infection less than 25%, score 2 for infection between 25% to 50%, score 3 for infection between 50%−75%, and score 4 for more than 75% involvement. In addition, the severity level [20] assigned as mild if the score is 1 or 2, moderate for a score between 2 and 5, severe if the score is more than 4 and infection less than 90%, critical if the infection is more than 90%. The detailed overview of the Covid-MANet multi-task classification and infection segmentation system for COVID-19 diagnosis is shown in Fig. 6 . Firstly, a binary lung mask is generated for each class using an encoder-decoder lung segmentation model. Then the predicted lung mask is superimposed with the input image to zero out the region outside the lungs generating a lung segmented image. This segmented image is given to the classification model by discarding irrelevant regions outside the lungs based on segmentation-based cropping. The classification model predicts the image into one pathology class based on higher output probability. For infection segmentation, the original image corresponding to the predicted COVID-19 image is given to the infection segmentation model that predicts the infection mask. Then for COVID diagnosed cases, quantification of the infection region is highlighted when the predicted infection mask and lung mask are superimposed with the original image. The infection masks are predicted at a threshold value of 0.5 and masked with corresponding lung masks to ensure infection regions outside the lungs are discarded. Accordingly, a COVID-19 positive sample is graded as mild, moderate, severe, and critical based on the proportion of pixels infected by the disease. Another case occurs when the model predicts the sample as COVID-19, but the infection segmentation system does not predict any disease pixel such a case is considered an Asymptomatic COVID-19 infection case. In addition, for each predicted class, an explainable diagnosis is achieved by generating the localization map focused by Grad-CAM [29] showing the region focused by model for classification.
Fig. 6

The framework of Covid-MANet methodology for classification and infection segmentation of COVID-19.

The framework of Covid-MANet methodology for classification and infection segmentation of COVID-19.

Segmentation loss function

The loss function is one of the useful components in the training of segmentation and classification models. The primary objective of our study is to differentiate COVID-19 from other lung diseases and localize the COVID-19 infection region. The primary focus is on the lungs region, so firstly segmentation loss aims to learn the lung segmentation task and secondly infection region segmentation task. Because of higher class imbalance, we developed a hybrid loss function suitable for lung as well as infection segmentation. The segmentation loss is computed as the mean of dice loss and the intersection over union loss. Both dice coefficient and intersection over union almost measure the overlapping of predicted mask and ground truth whereas the segmentation loss function computes the error. The objective of lung and infection segmentation models is to minimize segmentation loss by learning differentiation between foreground and background class pixels. Corresponding to lung segmentation task classes belongs to {lung, background} and for COVID-19 disease segmentation these are {disease lesion, background} [34] . Moreover, segmentation loss for lung and infection segmentation models in the forward pass is computed as the mean of dice and IoU loss using Eq. (1) . The dice and intersection over union loss are computed using Eq. (2) and Eq. (3) [45], respectively. Here yi and p(yi) represents ground truth and predicted mask output by the final layer. In addition, the loss function is updated in subsequent epochs by the backpropagation and gradient descent process.

Classification loss function

The classification dataset has a large class imbalance that may affect the results of the minority class. To eliminate this problem classification models are trained using a weighted cross-entropy loss function instead of simple cross-entropy loss. The weighted loss function alters weights assigned to each training example while computation of loss. If classes are balanced, every sample contributes equally to loss function but based on importance minority class samples are assigned more weights that show significant effect in training. In our case, we consider an equal contribution to all the classes. These are balanced by assigning new weights to all the classes such that positive and negative examples in each class contribute equally to the loss function computed using Eq. (4) [46] . Finally, these weights are fused to the classification loss that reduce false positives of the minority class. The multi-label classification loss function used for the training of the classification model is computed using Eq. (5).

Experimental details

In this section, we summarize detailed specifications of datasets used for lung segmentation, infection segmentation and classification of COVID-19 from other lung diseases. In addition, implementation details concerning model development, environment setup, and evaluation measures used for segmentation and diagnosis have been discussed in detail.

Dataset description and setup

This study aims to develop an automated deep learning model that assists radiologists in the early diagnosis and infection region quantification of COVID-19. To accomplish this study, we compiled well-known publicly available datasets following proper preprocessing guidelines established and used in almost all the research studies for the segmentation and classification of COVID-19. The classification dataset is compiled by merging covid-chestxray-dataset, chest-xray-pneumonia, and Tuberculosis ChestXrayImage DataSets for distinguishing COVID-19 from normal and abnormal lungs having Tuberculosis, viral/bacterial pneumonia. For infection segmentation, QaTa-COV19 dataset is used whereas detailed characteristics and references of these publicly available datasets are summarized as follows:

Lung and infection segmentation dataset

Table 1 specifies detailed resources of lung and infection segmentation datasets. The supervision dataset for lung segmentation contains a corpus of 7303 image masks pairs, compiled by merging five datasets corresponding to three classes only. The compiled dataset includes a corpus of 247 PA view images from the JSRT [47] dataset and their corresponding lung masks in the SCR [48] database. Montgomery Country and Shenzhen No.3 People's Hospital [49] datasets having a corpus of 138 PA view, 566 chest radiographs with their corresponding lung masks of TB, and normal classes are used. These are created by the U.S. National Library of Medicine (USNLM), Maryland, USA in collaboration with the Department of Health and Human Services and Shenzhen No.3 People's Hospital [49] at Guangdong Medical College in Shenzhen, China. For COVID-19, Chest X-ray-dataset created by Cohen et al. [50] is utilized having a corpus of 202 lung segmented image-mask pairs available publicly. Also, we consider 6150 pseudo lung masks of the RANZCR dataset [51] with almost similar UID compared to BP and VP cases. Supervision with this dataset enables better lung segmentation of unseen VP and BP in addition to other classes.
Table 1

Summary of segmentation dataset resources.

DatasetClasses#Images
3 JSRT [47] / SCR [48]Normal/Nodule247
2 Montgomery [49]Normal/TB138
2 Shenzhen [49]Normal/TB566
1 Cohen et al. [50]COVID-19202
4 RANZCR clip [51]6150
Total7303
5 QaTa-COV19 [34]COVID-19 Infection segmentation2951
Summary of segmentation dataset resources. In a real scenario, it is desirable to find the COVID-19 infection region and severity of the disease. To supervise the model for infection region segmentation, we utilized QaTa-COV19-v1[34] dataset compiled by Tampere University and Qatar University. The dataset comprises 2951 CXR images and corresponding ground truth masks for COVID-19 infection out of 4603 total COVID samples.

Classification dataset

Table 2 specifies detailed resources of classification datasets organized into five classes; COVID-19, normal, VP, BP, and TB. Specifically, a corpus of 476 COVID-19 CXR images of PA, AP, and AP supine view are collected from a similar dataset created by Cohen et al. [50]. For Comparison of COVID-19 with other lung diseases such as TB, VP, BP, and abnormal samples are collected from references [49,52]. Normal and pneumonia samples having a corpus of 1583 and 4273 images, taken from chest-xray-pneumonia dataset [52]. Because of the similarity between viral pneumonia and COVID-19, more specific categorization of pneumonia into 1483 viral and 2780 bacterial pneumonia are considered. In addition, more normal samples are taken from the Montgomery and Shenzhen dataset with a total sum of 1986 images. The tuberculosis samples are taken from reference with a corpus of 394 samples.
Table 2

Summary of classification dataset resources.

Dataset usedClasses# Of imagesTotal
1 Cohen et al. [50]COVID-19476476
6 Kermany et al. [52]BP27802780
VP14931493
Normal15831986
2 Montgomery [49]Normal80
2 Shenzhen [49]Normal323
2 Montgomery [49]TB58394
2 Shenzhen [49]TB336
Total7129
Summary of classification dataset resources.

Implementation setup

The implementation details for segmentation and classification networks summarized in this section. The lung segmentation task is accomplished using UNet and its two variants. These variants use ResNet and DenseNet blocks in place of two convolution layers of UNet. Each model is trained for a minimum of 70 epochs with a batch size of 16. The loss function is minimized using adam optimizer [53] with an initial learning rate of 0.00001 and an early stopping strategy. The learning rate is reduced by a factor of 10 if the loss is not reduced for 5 consecutive epochs. Models trained with input size 320 × 320 × 1 and the dataset split for training and testing is 90% and 10%, respectively. The classification model classifies input into one of five classes based on maximum predicted probability by a softmax activation function. The classification models trained for a maximum of 50 backpropagation epochs with a mini-batch size of 32. Again, adam optimizer with an initial learning rate of 0.00001 is used to minimize a weighted cross-entropy loss function. This supports an early stopping strategy with a patience value of 15 and the learning rate is reduced by a factor of 10 if the loss is not reduced for 7 consecutive epochs. The classification models are trained with input size 224 × 224 × 3 except for InceptionV3, which is trained with 299 × 299 × 3. In addition, the distribution statistics of the dataset into the training, testing, and validation set follow the guidelines to ensure that no patient overlap exists in the train, test, and validation set. The distribution of each class into training, testing, and validation set is specified in Table 3 . However, normal, BP, and VP classes were already labeled into training and testing set but the validation set contains only 8 samples. To ensure the validity of the model, we increase the validation set by splitting training samples into a validation set.
Table 3

Distribution of images for classification in five infection types.

SetCOVID-19NormalVPBPTB
Training354158612202348311
Testing8023314824245
Validation4217012519038
Total476198614932780394
Training: 5816, Validation: 565, Testing: 748, Total: 7129
Distribution of images for classification in five infection types. The infection segmentation network is trained on QaTa-Cov19 dataset to localize the COVID-19 infection region. Models trained with an input size of 224 × 224 × 3 for 50 backpropagations epochs and a mini-batch size of 4. The dataset split into the proportion of 80–20% for training and testing, whereas 10% of the training dataset is used for validation. Similar segmentation loss and adam optimizer is used for training with an initial learning rate of 0.0001. All these tasks are implemented in python using NVIDIA Tesla P100 GPU. Other libraries used for implementation include Keras, TensorFlow, and scikit-image.

Statistical measures

We statistically analyzed the segmentation and classification models based on metrics computed using a confusion matrix. Lung and infection segmentation models follow pixel-level classification where negative class corresponds to the background and positive class related to the lung or infection pixel. Similarly, classification models based on sample statistics where COVID-19 classified samples are considered as positive and non-COVID as a negative class [56]. Specifically, lung and infection segmentation models are evaluated using dice similarity coefficient, Intersection over union (IoU), Sensitivity, and Precision computed [14] as per equations; (6–10). Classification models are evaluated by Accuracy, Recall, Precision, and F1-score computed as per equations; (6,9,10, and 11). Moreover, COVID-19 sensitivity and COVID-19 precision are also taken into consideration since the correct classification of COVID-19 infected samples are more desirable. These are computed in the same way as recall and precision computed for a particular class. In short, accuracy refers to the ratio of correctly classified samples/pixels to the total corpus of samples/pixels. IoU and DSC quantitatively measure the overlap between predicted lung/infection segmentation masks and ground-truth masks. However, the only difference is that the latter one (DSC) considers the double weight of true positive pixels. The sensitivity refers to the rate of correctly classified positive class samples to the total number of positive class samples. The precision (PPV) refers to the proportion of correctly classified positive class samples to all the samples classified as positive. F1-score computes the harmonic mean of recall and precision. The segmentation model having a higher dice similarity coefficient and IoU score is considered as the best model. Similarly, for classification model higher accuracy, sensitivity, and COVID-19 sensitivity is desirable.

Results and analysis

In this section, we discuss results attained in all three tasks of the Covid-MANet network i.e., lung segmentation, disease recognition, and disease segmentation. The model development process performed an extensive number of experiments for improving the interpretation and recognition based on lung localization regions. The sensitivity analysis of each class has been discussed in search of a suitable model for each class separately. In addition, infection severity assessment and model robustness are cross-validated on an unseen dataset. Finally, the Grad-CAM technique is used to analyze the interpretation and explainability of models.

Results for segmentation

The Covid-MANet network accomplishes the primary aim of disease classification by analyzing segmented lungs. The performance achieved by the lung localization network with and without the involvement of RANZCR dataset on the test subset is shown in Table 4 . Recall that no public dataset has annotated masks of viral and bacterial pneumonia classes. The majority of previous studies analyzed segmentation for three classes by Shenzhen CXR and Montgomery dataset masks. The addition of RANZCR dataset enables better lung localization of classification data not only for viral and bacterial pneumonia classes but upright for all the classes. Among all models, it has been seen that dense connectivity of encoder blocks outperformed other blocks because of decisions based on the collective knowledge from previous layers. DenseUNet holds a leading position in the lung localization task with a dice score of 96.70 and IoU of 93.64%. The quantitative results indicate improvement in the overlapping of ground truth and predicted lung mask by 2.20% with the pseudo-generated RANZCR dataset. The output of lung localization models is compared in Fig. 7 with ground truth segmentation masks. Results indicate model reliably localize lungs region of different severity levels of COVID-19 including mild, moderate, and critical in addition to non-COVID-19 cases. However, the challenge with previous approaches lacks segmentation of critical and severe cases accurately because Montgomery and Shenzhen dataset comprises TB and the normal cases of medium and high quality only. Our compiled benchmark dataset benefits to overcome the challenge of lung localization for viral, bacterial, and low-quality COVID-19 pneumonia cases. Post-processed lung segmentation mask is masked with the corresponding image to generate a lung localization map ready for disease recognition by the classification model of the Covid-MANet network.
Table 4

Lung segmentation networks performance.

ModelDice.IOURecallPPV
Without RANZCRUNet94.5189.6292.8296.30
ResUNet95.2090.8994.8195.62
DenseUNet95.4991.4094.2696.78
With RANZCRUNet96.0492.4095.0097.14
ResUNet96.2992.8694.5698.15
DenseUNet96.7093.6495.6297.82
Fig. 7

Comparison of ground truth and predicted lung masks by the lung segmentation models.

Lung segmentation networks performance. Comparison of ground truth and predicted lung masks by the lung segmentation models.

Results of classification

The performance comparison of state-of-the-art classification models with the proposed model in each experiment is shown in Table 5 . In the exploration of the most coherent approach, different model variants are evaluated on an independent test set, where the model selection process relies on accuracy, F1-score, and Covid-19 sensitivity. Recall that the proposed multi-scale hybrid classification model MA-DenseNet201 is compared with eight other deep learning models. Image enhancement and attention mechanism are evaluated in combination with comparative deep learning models. The light attention module CBAM adds 0.9 M additional trainable parameters to these networks.
Table 5

Comparison of proposed model performance with state-of-the-art variants in considered experiments.

Experiment no.ModelAccuracy (%)Recall (%)Precision (%)F1-score (%)COVID-19 Sen. (%)COVID-19 (PPV) (%)
Experiment 1ResNet5076.3476.3478.4175.9377.5086.11
VGG1692.7892.7893.0392.7696.2597.45
VGG1993.8593.8593.9393.8897.5098.73
NASNetMobile90.9190.9191.0590.7697.5097.50
MobileNet92.5192.5192.7492.3898.7598.75
InceptionV390.6490.6490.8290.6096.2596.25
DenseNet12193.7293.7293.9293.6798.75100
DenseNet20190.7890.7891.0590.6998.7598.75
MA-DenseNet20191.8491.8492.1691.6998.7598.75
Experiment 2ResNet50 (S)87.8387.8388.1887.6880.0094.12
ResNet50 (CBAM)88.2488.2488.8688.0283.7594.37
ResNet50 (GC)89.0489.0489.2888.9187.5095.89
VGG1691.4491.4492.4791.6596.2593.90
VGG19 (S)92.1192.1192.3492.1992.5097.37
VGG19 (CBAM)91.8491.8492.3691.9995.0096.20
VGG19 (GC)92.5192.5192.7992.6095.0096.20
NASNetMobile89.5789.5790.4989.8492.50100
MobileNet92.9192.9193.6193.0793.7597.40
MobileNet (CBAM)92.2592.2593.0292.4295.0097.44
MobileNet (GC)92.5192.5192.7292.5793.7593.75
InceptionV388.7788.7789.9589.0596.2596.25
InceptionV3 (CBAM)89.7189.7190.7989.9695.0097.44
InceptionV3 (GC)89.5789.5789.9789.6993.7597.40
DenseNet12191.8491.8492.6591.999598.70
DenseNet2090.9190.9191.8691.1396.2595.06
DenseNet201 (CBAM)91.7191.7192.0891.8395.0097.44
DenseNet201 (GC)91.7191.7192.5191.9196.2597.47
MA-DenseNet20193.4593.4594.0793.5997.5098.70
Experiment 3ResNet5086.2386.2386.6989.9577.5095.38
ResNet50 (GC)87.8387.8388.0787.7783.7593.06
ResNet50 (TVF + GC)88.9088.9089.2288.9682.5095.65
ResNet50 (GC + CBAM)88.3788.3788.5888.0486.2595.83
VGG1689.1789.1790.3589.3896.2597.47
VGG1990.1190.1190.7990.2892.5093.67
VGG19 (GC)90.2490.2490.9090.4196.2593.90
VGG19 (TVF + GC)90.3790.3791.1690.5792.5096.10
VGG19 (GC + CBAM)89.8489.8490.4390.0093.7596.15
NASNetMobile87.8387.8388.6688.0893.7593.16
MobileNet90.7890.7891.6890.9995.0096.20
MobileNet (GC)90.7890.7891.4190.4595.0096.20
MobileNet (TVF + GC)90.6490.6491.1090.7893.7596.15
MobileNet (GC + CBAM)90.7890.7891.1890.9095.0095.00
InceptionV390.7890.7891.5090.959597.44
InceptionV3 (GC)90.7890.7891.1290.899597.44
InceptionV3 (TVF + GC)90.9190.9191.0790.9793.7597.40
InceptionV3 (GC + CBAM)90.5190.5190.8090.6193.7596.15
DenseNet12191.8491.8492.2491.9796.2596.25
DenseNet20192.1192.1192.3192.1893.7597.40
DenseNet201 (GC)91.9891.9892.1592.0495.0096.20
DenseNet201 (TVF + GC)91.8491.8492.0891.9395.0097.44
DenseNet201(GC+ CBAM)92.3892.3892.6892.4795.0096.20
MA-DenseNet20192.7892.7893.6792.9597.5097.50

*Italics text represents best model; GC stands for Gamma correction and TVF stands for Total variation filter.

Comparison of proposed model performance with state-of-the-art variants in considered experiments. *Italics text represents best model; GC stands for Gamma correction and TVF stands for Total variation filter. The performance of comparative models in experiment 1 along with the proposed model on whole slide images indicates VGG19 has higher accuracy and F1 score. VGG19 having an accuracy of 93.85% is followed by DenseNet121 and MobileNet with 92.72% and 92.51% scores. Despite having higher accuracy values, VGG19 has less interpretation and COVID-19 sensitivity, whereas MobileNet, DenseNet121, and the proposed model have higher sensitivity for COVID-19. In addition, the layers of the VGG19 model offer less confident analysis with low accuracy for COVID-19 cases at higher threshold values compared to the proposed MobileNet model as shown in Fig. 10. The MA-DenseNet201 model is better for COVID-19 in this experiment but confusion still exists between viral and bacterial pneumonia cases.
Fig. 10

The sensitivity comparison of models where first row shows result of class-wise sensitivity analysis by each of model in 1st,2nd and 3rd experiment and the second row shows sensitivity comparison at different confidence threshold values. The red marked bracket indicates the proposed model is better in all cases after lung localization approach.

Considering experiment 2, the performance and interpretation of the proposed model improved by classification based on lung localization regions. The performance measures of the DenseNet, MobileNet, and ResNet50 model improved after localization of lungs considering more reliable features for diagnosis whereas the VGG model decreases. Moreover, the proposed MA-DenseNet201 model outperformed all other models in terms of accuracy, F1 score, and COVID sensitivity with a score of 93.45%, 93.59%, and 97.50%, respectively. MA-DenseNet201 model is followed by the MobileNet model with accuracy and F1 score of 92.91% and 92.57%. Class-wise sensitivity and accuracy comparison at different threshold values insights that the proposed model is reliable even at higher threshold values. Compared to the previous experiment, the proposed model with a lung localization network enables reliable classification much focused on lung segmented regions and reduces the confusion between viral and bacterial pneumonia cases by improving the sensitivity of viral pneumonia. To improve interpretation, we proposed experimental analysis based on segmentation-based cropping, discarding regions outside segmented lungs. Similar to the previous experiment, the MA-DenseNet201 outperformed all other model variants with accuracy, F1 score, and COVID-19 sensitivity of 92.78%, 92.95%, and 97.50%, respectively. The proposed model is followed by the DenseNet201 and DenseNet121 models with accuracy of 92.38 and 91.84%. Moreover, the sensitivity and accuracy of the Covid-MANet approach at various threshold values are higher compared to other state-of-the-art models. In addition, interpretation of the proposed approach is maximum among all the experiments indicating the model's robustness in recognition of mild, moderate, and severe cases. Fig. 8 shows the confusion matrix of the MA-DenseNet201 model. The training and validation growth of models corresponding to these experiments are also shown in the second row of this figure. The schematic representation of ROC curve indicates AUC of each class for the top four performing networks as shown in Fig. 9 . These results reveal that the proposed model attains better AUC for all the classes after lung localization. Moreover, the validation accuracy in experiment 1 ranges between 80 and 85%, whereas it ranges from 90 to 94% in the second and third experiments.
Fig. 8

The first row shows confusion matrix of proposed model and second row representing learning curves in each best model.

Fig. 9

Visualization of ROC map for top four performing models representing area under curve for each of class.

The first row shows confusion matrix of proposed model and second row representing learning curves in each best model. Visualization of ROC map for top four performing models representing area under curve for each of class.

Sensitivity and confidence-score analysis

In disease diagnosis, it is desirable to have higher sensitivity since it measures the proportion of correctly classified positive samples to ensure maximum cases of COVID-19 are predicted correctly. Fig. 10 shows the sensitivity of models corresponding to each class in three classification experiments. Sensitivity analysis helps to build an ensemble model based on class-varied accuracies of the best suitable model for each class. From the results of the classification scenario, we conclude that the proposed model has higher sensitivity for COVID-19, Normal, viral pneumonia, and tuberculosis ranging from 93 to 100%. But ResNet50 is better in recognition for bacterial pneumonia. The maximum sensitivity score of 96.69% is achieved for bacteria class by ResNet50, whereas the proposed model achieved 87% for BP. In addition, ResNet50 has the least sensitivity for other classes, whereas the proposed model attained 97.5% for COVID-19, 94.85% for normal, 93.92% for VP, and 100% for TB classes. The proposed model improves the sensitivity of VP cases after lung segmentation, thus reducing the risk of overlapping with COVID-19. The sensitivity comparison of models where first row shows result of class-wise sensitivity analysis by each of model in 1st,2nd and 3rd experiment and the second row shows sensitivity comparison at different confidence threshold values. The red marked bracket indicates the proposed model is better in all cases after lung localization approach. To further explore the robustness of the proposed model, we computed sensitivity and accuracy at different confidence threshold values of 0.5, 0.7, 0.85, and 0.95. The second row of Fig. 10 and Table 6 show the sensitivity and accuracy of experimental analysis at different confidence threshold intervals. Schematic representation insights that the sensitivity of the proposed model is stable even at higher threshold values as compared with other models. Similarly, from these observations, we conclude that the proposed model outperformed other comparative models after segmentation, but MobileNet better before segmentation.
Table 6

Accuracy comparison of models in each experiment at different confidence threshold values.

%age AccuracyThresholdResNet50VGG16VGG19NASNet MobileInception V3DenseNet121DenseNet201MobileNetProposed
Without Segmentation0.589.1797.1997.5496.4796.4497.4596.2597.0396.71
0.782.3596.0996.2596.1296.0195.9895.8296.7396.57
0.8580.4592.9491.7194.8393.9594.3095.5096.3996.25
0.958087.6285.8091.2289.9191.0494.0995.5095.05
With Segmentation0.595.7296.5597.0395.8895.8596.7396.6897.1697.37
0.795.4895.5895.8595.0595.7796.7196.6596.7697.35
0.8593.2693.0793.0493.3694.6295.4095.5096.3997.05
0.9589.3089.4188.4790.5892.4593.1294.0195.4296.76
Segmentation based cropping0.595.2695.7496.3694.9196.3696.7196.8496.3397.11
0.794.4994.8696.0193.8296.0196.1796.7696.4997.16
0.8593.2392.0895.3291.0695.3294.6595.8895.9097.08
0.9589.3387.5492.8887.3792.8892.0894.0394.5496.71
Accuracy comparison of models in each experiment at different confidence threshold values.

Results of covid infection segmentation

The results of the infection segmentation models are shown in Table 7 . The evaluation of the infection quantification model is tested in two configurations: serial and parallel scenarios. In the serial scheme, segmented lungs by lung localization model given to classification are directly fed into the infection segmentation model, if the sample is diagnosed as COVID-19. In the parallel scheme, an original CXR image of diagnosed COVID-19 sample is given to the infection segmentation model for quantification and severity assessment of COVID-19. UNet and UNet with DenseNet201 encoder evaluated on both the schemes. Results reveal that the parallel scheme has a better dice score of 81.86% and 85% compared to 80.84% and 83.86% of the serial scheme. Therefore, the parallel scenario is preferred to train and evaluate the remaining experiments in search of a coherent approach for infection segmentation. Similar to lung localization, the infection segmentation network achieves higher dice and IoU score in both scenarios with the DenseNet model as an encoder. The UNet model attains the highest performance when DenseNet121 is used as an encoder with a dice score of 86.15%. Besides this UNet and UNet with DenseNet201 encoder has dice scores of 81% and 85%, respectively. Fig. 11 shows the qualitative comparison of top-performing infection segmentation networks compared with ground truth infection masks. This reveals the robustness of UNet with DenseNet121 and DenseNet201 encoder networks in COVID-19 infection segmentation of mild, severe, moderate, and critical severity levels.
Table 7

Performance comparison of infection segmentation models.

ModelAcc.DiceIOURecallPPV
UNet94.8981.8669.5981.3582.75
Attention UNet94.5780.1967.9380.0581.72
UNet++94.3479.2866.6578.7281.44
R2UNet95.1682.1670.6982.6382.82
UNet++ + VGG1995.3683.0371.9083.6583.49
UNet + ResNet5095.7784.497485.1684.80
UNet + DenseNet12197.0186.1576.9486.2986.92
UNet + DenseNet20196.938575.6986.0186.94
Fig. 11

The qualitative comparison of ground truth masks to the masks predicted by infection segmentation models where column 1 shows ground truth infection mask and column 2–7 shows masks predicted by infection segmentation models, respectively.

Performance comparison of infection segmentation models. The qualitative comparison of ground truth masks to the masks predicted by infection segmentation models where column 1 shows ground truth infection mask and column 2–7 shows masks predicted by infection segmentation models, respectively.

Computational complexity analysis

In this section, we discuss the complexity analysis of lung and infection segmentation networks in terms of trainable parameters and inference time. The number of trainable parameters associated with lung segmentation models, UNet, ResUNet, and DenseUNet is 7.7 M, 10.9 M, and 25 M, respectively. The number of trainable parameters for infection segmentation models AttentionUNet, R2UNet, UNet++, UNet++ with VGG19 has 8.7 M, 26.2 M, 8.6 M, 23.4 M, whereas the inference time for each batch is 14 ms, 18 ms, 34 ms, 15 ms, 30 ms, respectively. Besides, UNet with encoder ResNet50, DenseNet121, and DenseNet201 has trainable parameters of 32.5 M, 12.1 M, 26.3 M, and inference time of 24 ms, 35 ms, 55 ms. The UNet with DenseNet201 backbone has the highest inference time of 14 ms per sample and ResNet50 has the highest trainable parameters of 32.5 M. The overall worst inference time for disease recognition and segmentation is 150 ms whereas the system can process multiple batches in a second.

Hybrid classification model

In the comparison of experimental analysis, we view that despite having higher accuracy values of experiment 1 and experiment 2, has the least interpretation and generalization. These models classify images by considering irrelevant features other than the lungs. But the segmentation-based cropping approach achieves comparable performance measures and higher explainability by making attention to relevant lung regions. Based on class-wise sensitivity analysis, we conclude that the Covid-MANet approach is best suitable for COVID-19, normal, TB, and virus classes compared to other models. It is seldom to have a model that is equally good for all five classes. Based on the class-level awareness, we make an ensemble model on a segmentation-based classification approach by combining the prediction ability of the proposed MA-DenseNet201 model with ResNet50 and MobileNet model. The best weights assigned to respective models are w1 = 0.1 (ResNet50), w2 = 0.2 (MobileNet) and w3 = 0.3 (MA-DenseNet201), respectively. Since, Resnet50 improves sensitivity for bacteria class, whereas MA-DenseNet201 model is better for COVID, normal, virus, and TB classes. However, the MobileNet model follows depthwise separable convolutions followed by pointwise convolutions helpful in a different set of feature extraction. The proposed Covid-MANet ensemble network attains accuracy, precision, sensitivity, F1-score, and COVID-19 sensitivity of 95.05%, 95.40%, 95.05%, 95.19%, and 98.75%, respectively. Table 8 shows the result of an ensemble approach improving the respective evaluation metrics for respective classes.
Table 8

Classification performance of Covid-MANet ensemble classification model.

ClassesRecallPrecisionF1-scoreSupport
COVID98.7597.5398.1580
Normal93.9998.6496.02233
VP96.6281.7188.54148
BP92.9799.1195.94242
TB10010010045
Weighted Avg95.0595.4095.19748
Classification performance of Covid-MANet ensemble classification model. From a global perspective, the proposed study aims to develop an end-to-end clinical system having higher interpretation in distinguishing COVID-19 cases from healthy and other lung diseases with similar symptoms. The Covid-MANet multi-task model achieves the desired goal by three tasks; Task 1: lung segmentation, Task 2: COVID-19 detection from healthy/other lung diseases, and Task 3: Infection segmentation. DenseUNet model is best selected for lung localization, whereas for infection segmentation UNet with DenseNet121 encoder outperformed other comparative models. The proposed classification model MA-DenseNet201 achieves better results even at higher confidence intervals with lung segmentation, whereas MobileNet is better without lung segmentation followed by the MA-DenseNet201 model. The interpretation and explainability are maximum with the segmentation-based cropping approach. Also, we observed that shallow CNN models have the least interpretation and explainability compared to deep CNN models. The preliminary class-wise sensitivity analysis shows that MA-DenseNet201 has higher sensitivity for COVID-19, normal, VP, and tuberculosis. But the recognition of bacterial pneumonia is better by the ResNet50 model. Finally, to avoid misclassification and overlapping, the ensemble Covid-MANet model is developed based on class varied sensitivity analysis achieving COVID-19 sensitivity of 98.75%. In addition, the infection quantification and severity assessment module assist radiologists with a better treatment plan based on the progression of the disease. The proposed Covid-MANet model ensures correct diagnostic interest to classify images based on relevant lung regions because of the dual attention mechanism. Since relevant information lies in the lungs, firstly lung localization and secondly segmentation-based cropping impose model attention more on lungs rather than background information. We ensure the robustness of our proposed model on the basics of evaluation measures, interpretation, ROC curve, learning curves, class-wise sensitivity analysis, and attention map visualization. The proposed model outperforms other comparative models in accuracy and sensitivity with maximum interpretation by a segmentation-based cropping approach. The ROC curve reveals similar results showing the maximum AUC achieved by the proposed model in the segmentation-based cropping approach. Class-wise sensitivity analysis shows better performance of the model for COVID-19, Viral, TB, and normal cases but ResNet50 better recognizes bacterial infection in all the experiments. The proposed model is highly stable having better sensitivity and accuracy even at a higher confidence interval of 85% and 95%. The learning curve shows that the proposed model reached an early stopping condition with maximum training in fewer epochs. The reason for maximum interpretation and explainability with the proposed model is because of its multi-scale deep feature extraction ability by the DenseNet201 model given relevant lung areas. In addition, the robustness and generalization of the proposed model are cross-validated in more detail on the unseen QaTa-COV19 dataset in Section 4.5.

External validation on unseen CXR images

Precisely, to demonstrate the generalization of the proposed system in a real-world scenario, additional results on the unseen QaTa-COV19 dataset has been reported. Initially, lung localization maps are generated on QaTa-COV19 dataset to report results with and without segmentation. The dataset used for the external validation contains 2951 CXR samples of COVID-19 with ground truth infection segmentation masks. Out of 2951 COVID infected samples, COVID-19 sensitivity on unseen plain CXR samples achieved by top-performing models such as MobileNet has 96.54% (2849 correctly classified), VGG19 has 96.98% (2862 correctly classified), DenseNet201 has 95.89% (2830 correctly classified) and MA-DenseNet201 has 91.83% (2710 correctly classified). With lung segmentation MobileNet achieves 96.13% (2837 correctly classified), VGG19 has 89.96% (2655 correctly classified), DenseNet201 has 95.83% (2828 correctly classified) and MA-DenseNet201 has 98.17% (2897 correctly classified) sensitivity. Finally, with lung localization and segmentation-based cropping, MobileNet achieves 97.15% (2867 correctly classified), VGG19 has 96.23% (2840 correctly classified), DenseNet201 has 97.25% (2870 correctly classified) and MA-DenseNet201 has 97.32% (2872 correctly classified) sensitivity. However, the ensemble model gives covid-19 sensitivity of 98.20%. The results on the cross-validation dataset show similar patterns after localization of lungs as seen in preliminary experiments. These interesting observations insight that even with a limited high-quality training dataset, the proposed model gives promising results on much larger test datasets. Moreover, variability of the dataset does not affect the results of the proposed methodology since decision making of the proposed methodology focused on relevant lung regions by lung localization scheme. In contrast, whole slide images always come with some processing artifacts such as marker lines, tissue folds, and uneven sectioning that results in out-of-focus regions. The training of models on these images will produce unexpected results when encountered in test images. The output of the infection segmentation network on the unseen classification dataset is shown in Fig. 12 . The infection diagnosis system predicts the given sample into one pathology class and quantifies the percentage of infection found in the left and right lung separately to assist the doctor in a better treatment plan. Finally, COVID-19 diagnosed sample is assigned a severity level such as; mild/moderate/severe or critical based on the score assigned according to the RALE scoring system. However, our proposed system not only classifies COVID-19 but quantifies the severity of COVID-19 infection, more useful in real-time clinical applications to understand the progression of the disease. Moreover, the proposed system can be used in real clinical practice, since the inference time in the worst case takes less than 150 ms per sample. This means number of samples can be tested in a second.
Fig. 12

The infection segmentation and severity grading results by the proposed model on COVID-19 classified samples graded as mild, moderate, severe, and critical, where the first row shows both lung and infection mask contours giving intuition for infection quantification.

The infection segmentation and severity grading results by the proposed model on COVID-19 classified samples graded as mild, moderate, severe, and critical, where the first row shows both lung and infection mask contours giving intuition for infection quantification. The proposed model successfully distinguishes COIVD-19 cases from healthy and other lung diseases having similar symptoms as COVID-19. The comparative results of existing state-of-the-art methods on our dataset split under a uniform approach as the proposed model with 85% and 95% confidence threshold values are shown in Table 9 . The proposed Covid-MANet system constitutes of four tasks; Task 1 (T1) is semantic lung segmentation, Task 2 (T2) COVID-19 detection with and without lung segmentation and segmentation-based cropping (SBC), Task 3 (T3) is infection quantification and severity assessment, and Task 4 (T4) as the multi-label loss to curtail class imbalance. Each task plays an important role in improving the behavior of the model. MA-DenseNet201 directly applied on the whole slide image (task 2) shows less accuracy and sensitivity. The performance measures improve by lung segmentation i.e., Task 1 (T1) and Task 2 (T2). The accuracy, sensitivity, and COVID-19 sensitivity of the MA-DenseNet201 model are higher than other SOTA models. The segmentation-based cropping (SBC) approach gives comparable accuracy and sensitivity values. But the major advantage of the proposed approach lies in interpretation improvement as explained in the next paragraph. In addition, parallel analysis of COVID-19 is based on the direct infection segmentation module (Task 1 and Task 3) where the sample is classified as COVID-19 if at least one pixel is classified as a disease. This task leads to more false positives and generalization errors where most VP, BP, and TB samples are diagnosed as COVID-19 with mild symptoms. This shows a certain similarity of COVID-19 infection to other lung diseases. Moreover, the proposed Covid-MANet model can better recognize the patterns of each disease type thus reducing false positives. We can say that lung segmentation and segmentation-based cropping play a pivotal role in improving the interpretation and explainability of models. This approach adds more weights to lung regions thus providing stable results not affected by the variability of the dataset as mentioned in Section 4.5. The infection segmentation model finds the percentage proportion of infection and assigns severity levels showing the progression of the disease to radiologists. Multi-label loss reduces biasness with minority class COVID-19 because of imbalance and improves the sensitivity of minority class.
Table 9

Accuracy and sensitivity comparison of proposed model with existing SOTA approaches.

ModelAccuracy@Y%
Sensitivity@Y%
COVID-19 Sen.@Y%
Y = 95Y = 85Y = 95Y = 85Y = 95Y = 85
VGG19 + FC [27]92.8895.3244.3868.7158.876.2
ResNet50+FC [20]89.3393.2347.3268.0458.876.2
InceptionV3+FC [26]92.8895.3265.2479.146078.7
MobileNet+ FC [37]94.5495.9077.683.569093.8
DenseNet121+FC [44]92.0894.6561.4975.9381.286.2
DenseNet201+FC [26]94.0395.8871.7281.8186.292.5
MA-DenseNet201+T295.0596.2578.4785.5697.597.5
MA-DenseNet201+T1 + T296.7697.0586.0989.396.597.5
MA-DenseNet201+T1+ T2 + SBC96.7197.0886.36909596.5
Accuracy and sensitivity comparison of proposed model with existing SOTA approaches. The Grad-CAM analysis shows that the model focuses more on relevant lung regions for the classification of samples in each class after segmentation-based cropping. After a critical analysis, MobileNet is considered the best model in experiment 1. It achieves better performance measures at higher threshold values as compared to VGG19. Considering experiment 2 with lung localization network, MA-DenseNet201 outperforms other comparative models. In experiment 3, with segmentation-based cropping MA-DenseNet201 is considered the best model based on higher sensitivity, accuracy, and interpretation analysis. The quantitative and qualitative analysis for the COVID-19 class shows MobileNet gives 42% correct interpretation with whole slide images. But the interpretation improved by MA-DenseNet201 with lung segmentation network (Task 1 and Task 2) to 74%, which was 72% with MobileNet. The maximum interpretation achieved by the MA-DenseNet201 model is 96% after joint analysis by lung localization and segmentation-based cropping in experiment 3. Also, the multi-label classification loss function improves covid-19 sensitivity from 93.75% to 97.15%. Fig. 13 shows the infection map and localization map generated for each disease type by the Grad-CAM visualization tool, explaining the visual decisions focused by the classification model. For COVID-19 disease, the Grad-CAM map correlates with the infection segmentation map. The localization map shows that Covid-MANet focuses more on the relevant lungs region for the classification of each disease type. The proposed model can be useful in real clinical scenarios since the annotation of lungs, disease, and understanding the progression of the disease is a challenging and time-consuming task.
Fig. 13

Interpretation of disease map focused using Grad-CAM for each disease type by Covid-MANet improved after supervision by segmentation-based cropping.

Interpretation of disease map focused using Grad-CAM for each disease type by Covid-MANet improved after supervision by segmentation-based cropping.

Conclusions and future work

In the state of exponential reproduction rate of the COVID-19 pandemic, the only solution to lessen its growth rate is to perform mass screening, early diagnosis, isolation, and correct treatment plan. However, several countries face a shortage of testing kits, laboratories, and medical professionals to deal with the situation within time. At that difficult time, researchers around the world presented CXR as the standard screening tool for findings of COVID-19 in a time-efficient and cost-effective way. In this paper, we present an “explainable solution” for the recognition of COIVD-19 pneumonia from other common lung diseases using CXR radiographs. The proposed Covid-MANet model not only classifies disease but also quantifies infection that intuitions the progression of disease in the lungs. This is the most generic three-stage classification model developed so far works for five classes involving lung segmentation, COVID-19 detection, infection quantification, and severity assessment. The experiments conducted give promising results showing the robustness of the proposed model over existing studies on CXR radiographs. The Covid-MANet is a highly interpretable and generalizable model that benefits from lung segmentation and segmentation-based cropping. Moreover, internal and external validation reveals that the proposed methodology is more stable, and not affected by the variability of the dataset since model decision-making focused on relevant lung regions. In contrast, the results on the whole slide images lack generalization and interpretability because of some processing artifacts such as marker lines, uneven sectioning, and tissue folds that result in focus out-of-lungs. The advantage of the proposed methodology lies in following the real scenario of classification and infection segmentation of COVID-19 by recognition among five classes. Some studies directly classify COVID-19 based on infection detected by segmentation that works fine for healthy/COVID-19 cases but lacks when VP, BP, and TB samples having similar symptoms as COVID-19 occurs. Precisely, Covid-MANet can be applicable for clinical applications. The features of the proposed methodology are lung segmentation, proper documentation of pre-processing, training parameters, loss functions for reproducibility, explainable five-class classification model, higher interpretation, generalization, quantification, and severity grading as mild, moderate, severe, and critical based on RALE scoring system. We believe that the promising results of the proposed Covid-MANet system can guide radiologists in understanding the progression of the disease and better treatment plans in the early stages. Moreover, the Covid-MANet model can be deployed at airports, bus stands, railway stations, etc., for screening of large masses in less time. In future work, we will extend our proposed framework to modalities other than CXR images such as ultrasound and CT images for explainable diagnosis and localization of COVID-19 infection. We believe that the proposed model is beneficial in improving the sensitivity of COVID-19 with these modalities. In addition, to extend work on CXR images, we plan to increase the corpus of the COVID-19 class by collecting samples from different institutions and exploring more public repositories. Also, we annotate VP and BP samples and increase the corpus of annotated COVID-19 samples under the guidance of radiological experts.

Ethical approval

This article does not contain any studies with human participants or animals performed by any of the authors.

Funding

There is no source of funding for this article.

Declaration of Competing Interest

The authors declare that they have no conflict of interest.
[50]1https://github.com/ieee8023/covid-chestxray-dataset
[49]2https://lhncbc.nlm.nih.gov/LHC-publications/pubs/TuberculosisChestXrayImageDataSets.html
[47]3http://db.jsrt.or.jp/eng.php
[51]4https://www.kaggle.com/raddar/ranzcr-clip-lung-contours
[34]5https://www.kaggle.com/aysendegerli/qatacov19-dataset
[52]6https://www.kaggle.com/paultimothymooney/chest-xray-pneumonia
  33 in total

1.  Development of a digital image database for chest radiographs with and without a lung nodule: receiver operating characteristic analysis of radiologists' detection of pulmonary nodules.

Authors:  J Shiraishi; S Katsuragawa; J Ikezoe; T Matsumoto; T Kobayashi; K Komatsu; M Matsui; H Fujita; Y Kodera; K Doi
Journal:  AJR Am J Roentgenol       Date:  2000-01       Impact factor: 3.959

2.  The Royal Australian and New Zealand College of Radiologists (RANZCR) relative value unit workload model, its limitations and the evolution to a safety, quality and performance framework.

Authors:  A Pitman; D N Jones; D Stuart; K Lloydhope; K Mallitt; P O'Rourke
Journal:  J Med Imaging Radiat Oncol       Date:  2009-10       Impact factor: 1.735

Review 3.  A scoping review of transfer learning research on medical image analysis using ImageNet.

Authors:  Mohammad Amin Morid; Alireza Borjali; Guilherme Del Fiol
Journal:  Comput Biol Med       Date:  2020-11-13       Impact factor: 4.589

4.  Deep Learning COVID-19 Features on CXR Using Limited Training Data Sets.

Authors:  Yujin Oh; Sangjoon Park; Jong Chul Ye
Journal:  IEEE Trans Med Imaging       Date:  2020-05-08       Impact factor: 10.048

5.  BS-Net: Learning COVID-19 pneumonia severity on a large chest X-ray dataset.

Authors:  Alberto Signoroni; Mattia Savardi; Sergio Benini; Nicola Adami; Riccardo Leonardi; Paolo Gibellini; Filippo Vaccher; Marco Ravanelli; Andrea Borghesi; Roberto Maroldi; Davide Farina
Journal:  Med Image Anal       Date:  2021-03-31       Impact factor: 8.545

6.  A fuzzy-enhanced deep learning approach for early detection of Covid-19 pneumonia from portable chest X-ray images.

Authors:  Cosimo Ieracitano; Nadia Mammone; Mario Versaci; Giuseppe Varone; Abder-Rahman Ali; Antonio Armentano; Grazia Calabrese; Anna Ferrarelli; Lorena Turano; Carmela Tebala; Zain Hussain; Zakariya Sheikh; Aziz Sheikh; Giuseppe Sceni; Amir Hussain; Francesco Carlo Morabito
Journal:  Neurocomputing       Date:  2022-01-21       Impact factor: 5.719

7.  Clinical and Laboratory Approach to Diagnose COVID-19 Using Machine Learning.

Authors:  Krishnaraj Chadaga; Chinmay Chakraborty; Srikanth Prabhu; Shashikiran Umakanth; Vivekananda Bhat; Niranjana Sampathila
Journal:  Interdiscip Sci       Date:  2022-02-08       Impact factor: 2.233

8.  MetaCOVID: A Siamese neural network framework with contrastive loss for n-shot diagnosis of COVID-19 patients.

Authors:  Mohammad Shorfuzzaman; M Shamim Hossain
Journal:  Pattern Recognit       Date:  2020-10-17       Impact factor: 7.740

View more
  1 in total

1.  Image enhancement techniques on deep learning approaches for automated diagnosis of COVID-19 features using CXR images.

Authors:  Ajay Sharma; Pramod Kumar Mishra
Journal:  Multimed Tools Appl       Date:  2022-08-01       Impact factor: 2.577

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.