Literature DB >> 34276114

MIDCAN: A multiple input deep convolutional attention network for Covid-19 diagnosis based on chest CT and chest X-ray.

Yu-Dong Zhang1, Zheng Zhang2,3, Xin Zhang4, Shui-Hua Wang5.   

Abstract

BACKGROUND: COVID-19 has caused 3.34m deaths till 13/May/2021. It is now still causing confirmed cases and ongoing deaths every day.
METHOD: This study investigated whether fusing chest CT with chest X-ray can help improve the AI's diagnosis performance. Data harmonization is employed to make a homogeneous dataset. We create an end-to-end multiple-input deep convolutional attention network (MIDCAN) by using the convolutional block attention module (CBAM). One input of our model receives 3D chest CT image, and other input receives 2D X-ray image. Besides, multiple-way data augmentation is used to generate fake data on training set. Grad-CAM is used to give explainable heatmap.
RESULTS: The proposed MIDCAN achieves a sensitivity of 98.10±1.88%, a specificity of 97.95±2.26%, and an accuracy of 98.02±1.35%.
CONCLUSION: Our MIDCAN method provides better results than 8 state-of-the-art approaches. We demonstrate the using multiple modalities can achieve better results than individual modality. Also, we demonstrate that CBAM can help improve the diagnosis performance.
© 2021 Elsevier B.V. All rights reserved.

Entities:  

Keywords:  Automatic differentiation; COVID-19; Chest CT; Chest X-ray; Convolutional neural network; Data harmonization; Deep learning; Multimodality; Multiple input

Year:  2021        PMID: 34276114      PMCID: PMC8277963          DOI: 10.1016/j.patrec.2021.06.021

Source DB:  PubMed          Journal:  Pattern Recognit Lett        ISSN: 0167-8655            Impact factor:   3.756


Introduction

COVID-19 (also known as coronavirus) pandemic is an ongoing infectious disease caused by severe acute respiratory syndrome (SARS) coronavirus 2 [1]. As of 13/May/2021, there are over 161.14m confirmed cases and over 3.34m deaths attributed to COVID-19. The cumulative deaths of the top 10 countries are shown in Fig. 1 .
Fig. 1

Top 10 countries in terms of cumulative deaths (13/May/2021).

Top 10 countries in terms of cumulative deaths (13/May/2021). The main symptoms of COVID-19 are a low fever, a new and ongoing cough, a loss or change to taste and smell. In UK, the vaccines approved were developed by Pfizer/BioNTech, Oxford/AstraZeneca, and Moderna. The joint committee on vaccination and immunization (JCVI) [2] determines the order in which people will be offered the vaccine. At April/2021, people aged 50 and over, people of clinically (extremely) vulnerable, people living or working in the care homes, and health care providers, people with a learning disability are being offered. Two COVID-19 diagnosis methods are available. The first method is viral testing to test the existence of viral RNA fragments [3]. The shortcomings of swab test [4] are two folds: (i) the swab samples may be contaminated and (ii) it needs to wait from several hours to several days to get the results. The other method is chest imaging. There are two main different chest imaging available: chest computed tomography (CCT) and chest X-ray (CXR) CCT is one of the best chest imaging techniques so far, because it provides the finest resolution and it is capable of recognizing extremely small nodules [5]. It provides high-quality volumetric 3D chest data. On the other hand, CXR performs poor on soft tissue contrast, and it only provides 2D image [6]. In this paper, we aim to fuse CCT and CXR images, and expects the fusion can improve the performance compared to using CCT or CXR individually. Besides, we create a novel multiple input deep convolutional attention network (MIDCAN) that can handle CCT and CXR images simultaneously, and present the diagnosis output. The contributions of this study are itemized briefly as following five points: Attention mechanism, convolutional block attention module, is included in the proposed MIDCAN model to improve the performance; The proposed MIDCAN model can handle CCT and CXR images simultaneously; Multiple-way data augmentation is employed to overcome overfitting problem; This proposed MIDCAN model gives more accurate performances than individual modality-based approaches; The proposed MIDCAN model is superior to state-of-the-art COVID-19 diagnosis approaches.

Literature survey

From previous year, AI field has investigated ongoing researches on automatic COVID-19 diagnosis, which can save the workloads of manual labelling. For the CCT image based COVID-19 diagnosis, Chen (2020) [7] employed gray-level occurrence matrix (GLCM) as feature extraction method. The authors then used support vector machine (SVM) as the classifier. Yao (2020) [8] combined wavelet entropy (WE) and biogeography-based optimization (BBO). Wu (2020) [9] presented a novel method—wavelet Renyi entropy (WRE) to help diagnose COVID-19. El-kenawy, Ibrahim (2020) [10] proposed a feature selection voting classifier (FSVC) approach for COVID-19 classification. Satapathy (2021) [11] combined DenseNet with optimization of transfer learning setting (OTLS). Saood and Hatem (2021) [12] explored two structurally-different deep learning (DL) methods—U-Net and SegNet—for COVID-19 CT image segmentation. On the other side, there are several successful AI models for CXR image based COVID-19 diagnosis. For example, Ismael and Sengur (2020) [13] presented a multi-resolution analysis (MRA) approach. Loey, Smarandache (2020) [14] combined generative adversarial network (GAN) with GoogleNet. Their method is abbreviated as GG. Togacar, Ergen (2020) [15] employed social mimic optimization (SMO) for feature selection and combination. Das, Ghosh (2021) [16] used weighted average ensembling technique with convolutional neural network (CNN) for automatic COVID-19 detection. The main shortcomings of above approaches are three points: (i) They only consider individual modality, either CCT or CXR. (ii) Their AI models are either traditional feature extraction plus classifier model, or modern deep neural network models. Nevertheless, their models lack attention mechanism. (iii) Efficient measures to resist overfitting are missing. To solve or alleviate above three shortcomings, we proposed the multiple input deep convolutional attention network. The dataset and details of our method will be discussed in Sections 3 and 4, respectively.

Dataset

Data harmonization

This retrospective study was granted to exempt ethical approval. 42 COVID-19 patients and 44 healthy controls (HCs) were recruited. All the data were collected from local hospitals. Each subject takes a CCT scan a CXR scan, and thus generate a CCT image and CXR image . Due to different chest sizes of different people and the different sources of scanning machines, the height of image and the size of vary. To make a homogeneous dataset, data harmonization [17] is used. The central 64 slices of CCT image and the central rectangle region of CXR image are reserved. The height and width of CCT slices are resized to , and the CXR image is resized to . They are named and . We choose 64 and 2048 because we find this can keep the lung part of images while get rid of unrelated body tissues. The details are displayed in Algorithm 1 .
Algorithm 1

Data harmonization.

InputCCT image A0(n) and CXR image B0(n) of subject n.
Step 1For CCT image A0(n)64 central slices are reserved, and top/bottom slices are removed
Step 2For CXR image B0(n)Central rectangle region is reserved, and outskirt pixels are removed.
Step 3CCT images are resized to 1024×1024, and CXR image is resized to 2048×2048.
OutputCCT image A1(n) and CXR image B1(n).size[A1(n)]=1024×1024×64.size[B1(n)]=2048×2048.
Data harmonization.

Data Preprocessing

Second, data preprocessing (See Fig. 2 ) is used since both CCT and CXR image contain redundant/unrelated spatial information and their sizes are still too large. First, all the CCT and CXR images are grayscaled. Second, histogram stretch is carried to enhance the image contrast, where stand for the minimum and maximum grayscale values of our images. Third, the margins at four directions are cropped (e.g., the text in the right side and the check-up bed in bottom side of CCT images, the neck part in the top side of CXR images, the background regions at four directions, etc.). Finally, CCT images are resized to and CXR images are resized to .
Fig. 2

Flowchart of preprocessing.

Flowchart of preprocessing. Fig. 3 gives the examples of preprocessed images of a COVID-19 patient. Fig. 3(a) displays one slice out of 16 CCT slices, and Fig. 3(b) displays the CXR image.
Fig. 3

Pre-processed images of one COVID-19 patient.

Pre-processed images of one COVID-19 patient.

Methodology

Convolutional block attention module

Table 1 gives the abbreviation list. DL has gained many successes in prediction/classification quests. Among all the DL structures, convolutional neural network (CNN) [18, 19] is particularly suitable for analyzing 2D/3D images. To help boost the performance of CNN, researches are proposed to modify CNN structures in terms of either depth, or cardinality, or width. Newly, scholars have studied on attention mechanism, and attempted to integrate attention to DL structures. For example, Hu, Shen (2020) [20] proposed squeeze-and-excitation (SE) network. Woo, Park (2018) [21] presented a new convolutional block attention module (CBAM), that improves the traditional convolutional block (CB) by integrating attention mechanism. This study we choose CBAM because CBAM can provider both spatial attention and channel attention, compared to SE.
Table 1

Abbreviation list.

MeaningsAbbreviations
AMactivation map
AIartificial intelligence
APaverage pooling
BNbatch normalization
CAMchannel attention module
CCTchest computed tomography
CXRchest X-ray
CBconvolutional block
CBAMconvolutional block attention module
CNNconvolutional neural network
DAdata augmentation
DLdeep learning
FMIFowlkes–Mallows index
MCCMatthews correlation coefficient
MPmax pooling
MSDmean and standard deviation
ReLUrectified linear unit
SAPNsalt-and-pepper noise
SAMspatial attention module
SNspeckle noise
SEsqueeze-and-excitation
Abbreviation list. Take a 2D-image input as an example, Fig. 4 (a) displays the structure of a traditional CB. The output of previous block was sent to -repetitions of convolution layer, batch normalization (BN), and rectified linear unit (ReLU) layer. Finally, the -repetitions is followed by a pooling layer. The output is named activation map (AM), symbolized as , where stands for the sizes of channel, height, and width, respectively.
Fig. 4

Structural comparison.

Structural comparison. In contrast to Fig. 4(a), Fig. 4(b) adds the structure of CBAM, by which two modules: channel attention module (CAM) and spatial attention module (SAM) are added to refine the activation map . Suppose the CBAM applies a 1D CAM and a 2D SAM in sequence to the input . Hence, the channel-refined activation map can be obtained as:And the final refined AMwhere means the element-wise multiplication. is the refined AM, which replaces the of traditional CB output, and it will be sent to the next block. Note if the above two operands are not with the same dimension, the values are reproduced so that (i) the spatial attentional values are copied along the channel dimension, and (ii) the channel attention values are copied along the spatial dimension.

Channel Attention Module

CAM is firstly defined. Both max pooling (MP) and average pooling (AP) are employed, making two features and as shown in Fig. 5 (a).
Fig. 5

Flowchart of two modules.

Flowchart of two modules. Both and are thenceforth sent to a shared multi-layer perceptron (MLP) to make the output AMs, that are then merged via element-wise summation . The merged sum is lastly forwarded to . That is,where is sigmoid function. To decrease the parameter space, the hidden size of MLP is fixed to , where stands for the reduction ratio. Assume and mean the MLP weights (See Fig. 5a), respectively, equation (4) can be rewritten as:Note and are shared by both and . Fig. 5(a) displays the diagram of CAM.

Spatial Attention Module

Next, SAM is defined in Fig. 5(b). The spatial attention module is a complementary procedure to the previous CAM . The average pooling and max pooling are harnessed again to the channel-refined activation map , Both and are two dimensional AMs: . They are concatenated via concatenation function together along the channel dimension as Afterwards, the concatenated AM is passed into a standard convolution with the size of , followed by sigmoid function . Overall, we attain: The is then element-wisely multiplied by to get the final refined AE . See Equation (2). The diagram of SAM is portrayed in Fig. 5(b).

Single Input and Multiple Input Deep Convolutional Attention Networks

In this study, we proposed a novel multiple-input deep convolutional attention network (MIDCAN) based on the ideas of CBAM and multiple-input. The structure of this proposed MIDCAN is determined by trial-and-error method. The variable in each block varies, and we found the best values are chosen in the range from . We tested values larger than 3, which increase the computation burden, but the performances do not increase. The structure of proposed shown below in Fig. 6 (a), which is composed of two inputs. The left input is “Input-CCT” where CCT images are passed into the network. The right input is “Input-CXR” where CXR images are passed into the network. Suppose and stand for the number of CBAM blocks individually. We set in this study by trial-and-error.
Fig. 6

Variables and sizes of AMs of three proposed models.

Variables and sizes of AMs of three proposed models. For the left branch, the CCT input goes through 3D-CBAMs and generates the output AM , which is then flattened into . Similarly, the CXR input at right branch goes through 2D-CBAMs, generates the output AM , which is flattened into . The deep CCT features and deep CXR features are then concatenated via concatenation function as Here note than in our experiments, we use ablation studies, where we set two models: single-input deep convolutional attention network (SIDCAN) models, which remove the left and right branches, respectively. The first SIDCAN model, shown in Fig. 6(b), will only use CCT features, i.e., This model is given a short name as SIDCAN-CCT. The second SIDCAN model will only use CXR features, i.e., This model is named as SIDCAN-CXR. Its flowchart is displayed in Fig. 6(c). Those two models will used as comparison method in our experiments. The feature is then passed to two fully-connected layers [22]. The first FCL contains 500 neurons, and the last FCL contains neurons, where stands for the number of classes. In this study . Finally, a softmax layer [23] turns the to probability. The loss function of this MIDCAN is cross entropy [24] function. Table 2 gives the details of proposed MIDCAN. For the kernel parameter in Table 2, [3 × 3 × 3, 16]x3, [/2/2/2] stands for 3 repetitions of 16 filters with each size of , following by a pooling with pooling factor of 2, 2, and 2 along three dimensions, respectively. In FCL stage, the kernel parameter gives the size of weight matrix and bias vector, respectively.
Table 2

Details of proposed MIDCAN model.

NameKernel ParameterVariable and size
Input-CCTsize(A2)=256×256×16
3D-CBAM-1[3 × 3 × 3, 16]x3, [/2/2/2]size(A3)=128×128×8×16
3D-CBAM-2[3 × 3 × 3, 32]x2, [/2/2/1]size(A4)=64×64×8×32
3D-CBAM-3[3 × 3 × 3, 32]x2, [/2/2/2]size(A5)=32×32×4×32
3D-CBAM-4[3 × 3 × 3, 64]x2, [/2/2/1]size(A6)=16×16×4×64
3D-CBAM-5[3 × 3 × 3, 64]x2, [/2/2/2]size(A7)=8×8×2×64
Flattensize(A8)=8192
Input-CXRsize(B2)=256×256
CBAM-1[3 × 3, 16]x3, [/2/2]size(B3)=128×128×16
CBAM-2[3 × 3, 32]x2, [/2/2]size(B4)=64×64×32
CBAM-3[3 × 3, 64]x2, [/2/2]size(B5)=32×32×64
CBAM-4[3 × 3, 64]x2, [/2/2]size(B6)=16×16×64
CBAM-5[3 × 3, 128]x2, [/2/2]size(B7)=8×8×128
Flattensize(B8)=8192
Concatenatesize(F1)=16,384
FCL-1500 × 16384, 500 × 1size(F2)=500
FCL-22 × 500, 2 × 1size(F3)=2
Softmax
Details of proposed MIDCAN model.

18-way data augmentation

Data augmentation (DA) [25] is an important utensil over the training set to avoid overfitting of classifiers when applied to test set. Meanwhile, DA can overcome the small-size dataset problem. Recently, Wang (2021) [26] proposed a novel 14-way data augmentation (DA), which used seven different DA techniques to the preprocessed training image and its horizontal mirrored image , respectively. Cheng (2021) [27] presented a 16-way DA, and used PatchShuffle technique to avoid overfitting. This study enhances the 14-way DA method [26] to 18-way DA, by adding two new DA methods: salt-and-pepper noise (SAPN) and speckle noise (SN) on both and . Use as an example, the SAPN altered image is symbolized as with its values are set aswhere stands for noise density, and the probability function. and correspond to black and white colors, respectively. On the other side, the SN altered image is defined aswhere is a uniformly distributed random noise, of which the mean and variance are symbolized as and , respectively. Take Fig. 3(b) as the example, Fig. 7 (a-b) display the SAPN and SN altered images, respectively. Due to the page limit, the results of other DA are not shown in this paper.
Fig. 7

Examples of newly proposed DA methods.

Examples of newly proposed DA methods. Let stands for the number of DA techniques to the preprocessed image , and stands for the number of new generated images for each DA. This proposed -way DA algorithm is a four-step algorithm depicted below: First, geometric/photometric/noise-injection DA transforms are utilized on preprocessed train image ,. We use to denote each DA operation. See each DA operations yields new images. Thus, for a given image , we yield different data set , and each dataset contains new images. Second, horizontally mirrored image is generated aswhere means horizontal mirror function. Third, all the DA methods are carried out on the mirror image , and generate different dataset . Four, the raw image , the horizontally mirrored image , all -way results of preprocessed image , and -way DA results of horizontally mirrored image , are fused together using concatenation function , as defined in Eq. (9). The final combined dataset is defined as Therefore, one image will generateimages (including the original image ). Note in our dataset, different will be assigned to CCT training images and CXR images since CCT images are 3D and CXR images are 2D. That means for each DA, we have new images for each CCT image and new images for each CXR image.

Implementation and evaluation

-fold cross validation is employed on both datasets. Suppose confusion matrix over r-th () run and k-th () fold is defined aswhere stand for TP, FN, FP, and TN, respectively. P stands for positive class, i.e., COVID-19, and N means negative class, i.e., healthy control. represents the index of trial/fold, and r stands for the index of run. At -th trial, the -th fold is used as test, and all the left folds are used as training, Note that is calculated based on each test fold, and are then summarized across all trials, as shown in Fig. 8 . Afterwards, we get the confusion matrix at r-th run as
Fig. 8

Diagram of one run of -fold cross validation.

Diagram of one run of -fold cross validation. Seven indicators are computed based on the confusion matrix over r-th run .where first four indicators are: sensitivity, specificity, precision, and accuracy. Those four indicators are commonly used. Their definitions can be found easily. is F1 score. is Matthews correlation coefficient (MCC)and is the Fowlkes–Mallows index (FMI). There are two indicators and using all the four basic measures . Considering the range of is , and the range of is , we finally choose as the most important indicator. Besides, Chicco, Totsch (2021) [28] stated that MCC is more reliable than many other indicators. Above procedure is one run of -fold cross validation. We run the -fold cross validation runs. The mean and standard deviation (MSD) of all seven indicators are calculated over all runs.where stands for the mean value, and stands for the standard deviation. The MSDs are reported in the format of .

Experiments, results, and discussions

Parameter setting

Table 3 itemizes the parameter setting. Here the minimum value and maximum value of our images are set to 0 and 255, respectively. The size of preprocessed CCT images and CXR images are set to and , respectively. The number of CBAM blocks for CCT and CXR branches are set to 5. The noise density of SAPN is set to 0.05. The mean and variance of uniform distributed noise in SN are set to 0 and 0.05, respectively. Nine different DA methods are used, so we have an 18-way DA if we consider both raw training image and its horizontal mirrored image. For each DA, 30 new images are generated for each CXR image, and 90 new images are generated for each CCT image. The number of -fold is set to . We run our model runs.
Table 3

Parameter setting.

ParameterValue
(vmin,v,max)(0, 255)
HCCT×WCCT×CCCT256×256×16
HCXR×WCXR256×256
NCCCT5
NCCXR5
γdsa0.05
Umsn0
Uvsn0.05
Ma9
MbCXR30
MbCCT90
K10
R10
Parameter setting.

Statistics of proposed MIDCAN

We use two modalities, CCT and CXR, in this experiment. The structure of our model is shown in Fig. 6(a). The statistical results of proposed MIDCAN are shown in Table 4 . As it shows, the sensitivity, specificity, precision and accuracy are 98.10±1.88, 97.95±2.26, 97.92±2.24, and 98.02±1.35, respectively. Moreover, the F1 score is 97.98±1.37, the MCC is 96.09±2.66, and FMI is 97.99±1.36.
Table 4

Statistical results of proposed MIDCAN model.

Runη1η2η3η4η5η6η7
197.62100.00100.0098.8498.8097.7098.80
297.6297.7397.6297.6797.6295.3597.62
3100.00100.00100.00100.00100.00100.00100.00
4100.0097.7397.6798.8498.8297.7098.83
595.2497.7397.5696.5196.3993.0496.39
6100.0093.1893.3396.5196.5593.2696.61
7100.00100.00100.00100.00100.00100.00100.00
897.6295.4595.3596.5196.4793.0596.48
995.24100.00100.0097.6797.5695.4497.59
1097.6297.7397.6297.6797.6295.3597.62
MSD98.10±1.8897.95±2.2697.92±2.2498.02±1.3597.98±1.3796.09±2.6697.99±1.36
Statistical results of proposed MIDCAN model.

Effect of multimodality and attention mechanism

We compare multiple-modality against single-modality. Two models, viz., SIDCAN-CCT and SIDCAN-CXR, shown in Fig. 6(b-c) are used. Meanwhile, using attention and not using attention are compared. The comparison results are shown in Table 5 , where NA means no attention. Fig. 9 presents the error bar comparison of all the six setting. Comparing using attention against without using attention, we can observe the attention mechanism does help improve the classification performance, which is coherent with the conclusion of Ref. [21].
Table 5

Comparison of different settings.

Methodη1η2η3η4η5η6η7
MIDCAN98.10±1.8897.95±2.2697.92±2.2498.02±1.3597.98±1.3796.09±2.6697.99±1.36
SIDCAN-CCT96.19±1.2395.91±1.4495.76±1.4096.05±0.6095.96±0.6092.11±1.2095.97±0.60
SIDCAN-CXR93.81±3.0193.86±2.6493.70±2.4893.84±0.7893.69±0.8587.78±1.5693.72±0.84
MIDCAN(NA)94.29±2.3094.32±1.6194.10±1.4594.30±0.8694.17±0.9488.65±1.6994.18±0.93
SIDCAN-CCT(NA)92.86±1.5993.64±2.0993.36±2.0293.26±0.7493.08±0.7186.55±1.4993.10±0.72
SIDCAN-CXR(NA)89.52±2.3090.68±2.9290.29±2.6390.12±0.8289.85±0.7680.31±1.6489.88±0.76
Fig. 9

Error bar comparison of six different settings.

Comparison of different settings. Error bar comparison of six different settings. Meanwhile, if we compare MIDCAN with two SIDCAN models, we can conclude that multimodality has the better performance than single modalities (both CT and CXR). Fig. 10 displays the ROC curves of six settings. The blue patch corresponds to the lower bound and upper bound. For the first three models using attention, we can observe their AUCs are 0.9855, 0.9695, and 0.9567 for MIDCAN, SIDCAN-CCT, and SIDCAN-CXR, respectively. If removing the CBAM module, we can observe from the bottom part of Fig. 10, that the corresponding AUCs decrease to 0.9512, 0.9361, and 0.9262, respectively. In addition, multimodality is proven to give better performance than using single-modality.
Fig. 10

ROC curves of six settings.

ROC curves of six settings.

Explainability of proposed model

Fig. 11 presents the manual delineation and heatmap results of Fig. 3. The heatmap images are generated via Grad-CAM method [29].
Fig. 11

Manual delineation and heatmap results of one patient.

Manual delineation and heatmap results of one patient. From Fig. 11, we can observe the proposed MIDCAN model is able to capture the lesions of both CCT image and CXR image accurately. This explainability via Grad-CAM can help the doctors, radiologists, and patients to better understand how our AI model works.

Comparison to State-of-the-art approaches

We compare the proposed method “MIDCAN” with 8 state-of-the-art methods: GLCM [7], WE-BBO [8], WRE [9], FSVC [10], OTLS [11], MRA [13], GG [14], SMO [15]. Those methods were carried out on single modality dataset depending on their original paper reported (either CCT or CXR), so we test those methods in the corresponding SIDCAN models and single-modality dataset. All the methods were evaluated via 10 runs of 10-fold cross validation. The MSD results of all approaches on ten runs of 10-fold cross validation are pictured in Fig. 12 , which sorts all the methods in terms of , and itemized in Table 6 .
Fig. 12

3D bar plot of approach comparison.

Table 6

Comparison with SOTA approaches (Unit: %).

Approachη1η2η3η4η5η6η7
GLCM [7]71.90±4.0278.18±3.8976.04±2.4175.12±0.9873.80±1.4950.35±1.9173.89±1.39
WE-BBO [8]74.05±4.8274.77±3.9373.83±1.8474.42±0.7873.81±1.6548.98±1.6573.88±1.67
WRE [9]86.43±3.1886.36±3.8686.01±3.1386.40±0.5686.12±0.3972.95±1.1586.17±0.40
FSVC [10]91.90±2.5690.00±2.4489.85±1.9990.93±0.4990.82±0.5581.97±0.9990.85±0.56
OTLS [11]95.95±2.2696.59±1.6196.45±1.5696.28±1.0796.17±1.1392.60±2.0996.19±1.12
MRA [13]86.43±3.9090.45±2.7989.71±2.6388.49±2.0887.98±2.2777.09±4.1788.02±2.26
GG [14]93.33±2.7090.00±4.4490.13±3.8191.63±1.5391.61±1.3583.49±2.8491.67±1.30
SMO [15]93.10±2.3795.23±2.5094.99±2.4594.19±1.1093.99±1.1388.45±2.1694.02±1.11
MIDCAN (Ours)98.10±1.8897.95±2.2697.92±2.2498.02±1.3597.98±1.3796.09±2.6697.99±1.36
3D bar plot of approach comparison. Comparison with SOTA approaches (Unit: %). From Table 6, we can observe that this proposed MIDCAN outperforms all the other 8 comparison baseline methods in terms of all indicators. The reason why our MIDCAN method is the best lie in following three facts: (i) we propose to use multiple modality instead of traditional single modality; (ii) CBAM is used in our network that attention mechanism can help our AI model focuses on the lesion region; (iii) multiple-way data augmentation is employed to overcome overfitting.

Conclusion

This paper proposed a novel multiple input deep convolutional attention network (MIDCAN) model for diagnosis of COVID-19. The results show our method achieves a sensitivity of 98.10±1.88%, a specificity of 97.95±2.26%, and an accuracy of 98.02±1.35%. In the future researches, we shall carry out several attempts: (i) expand our dataset; (ii) include other advanced network strategies, such as graph neural network; (iii) collect IoT signals of subjects.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  16 in total

1.  Squeeze-and-Excitation Networks.

Authors:  Jie Hu; Li Shen; Samuel Albanie; Gang Sun; Enhua Wu
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2019-04-29       Impact factor: 6.226

2.  Synthesis of brain tumor multicontrast MR images for improved data augmentation.

Authors:  Sunho Kim; Byungjai Kim; HyunWook Park
Journal:  Med Phys       Date:  2021-03-22       Impact factor: 4.071

3.  COVID-19 detection using deep learning models to exploit Social Mimic Optimization and structured chest X-ray images using fuzzy color and stacking approaches.

Authors:  Mesut Toğaçar; Burhan Ergen; Zafer Cömert
Journal:  Comput Biol Med       Date:  2020-05-06       Impact factor: 4.589

4.  Comparative evaluation of nasopharyngeal swab and saliva specimens for the molecular detection of SARS-CoV-2 RNA in Japanese patients with COVID-19.

Authors:  Daisuke Sakanashi; Nobuhiro Asai; Akiko Nakamura; Narimi Miyazaki; Yuzuka Kawamoto; Tomoko Ohno; Atsuko Yamada; Isao Koita; Hiroyuki Suematsu; Mao Hagihara; Arufumi Shiota; Ai Kurumiya; Miki Sakata; Syunji Kato; Yuki Muramatsu; Yusuke Koizumi; Takaaki Kishino; Wataru Ohashi; Yuka Yamagishi; Hiroshige Mikamo
Journal:  J Infect Chemother       Date:  2020-09-30       Impact factor: 2.211

5.  COVID-19 lung CT image segmentation using deep learning methods: U-Net versus SegNet.

Authors:  Adnan Saood; Iyad Hatem
Journal:  BMC Med Imaging       Date:  2021-02-09       Impact factor: 1.930

6.  Determinants of mortality in a large group of hemodialysis patients hospitalized for COVID-19.

Authors:  Kenan Turgutalp; Savas Ozturk; Mustafa Arici; Necmi Eren; Numan Gorgulu; Mahmut Islam; Sami Uzun; Tamer Sakaci; Zeki Aydin; Erkan Sengul; Bulent Demirelli; Yavuz Ayar; Mehmet Riza Altiparmak; Savas Sipahi; Ilay Berke Mentes; Tuba Elif Ozler; Ebru Gok Oguz; Bulent Huddam; Ender Hur; Rumeyza Kazancioglu; Ozkan Gungor; Bulent Tokgoz; Halil Zeki Tonbul; Alaattin Yildiz; Siren Sezer; Ali Riza Odabas; Kenan Ates
Journal:  BMC Nephrol       Date:  2021-01-14       Impact factor: 2.388

7.  PSSPNN: PatchShuffle Stochastic Pooling Neural Network for an Explainable Diagnosis of COVID-19 with Multiple-Way Data Augmentation.

Authors:  Shui-Hua Wang; Yin Zhang; Xiaochun Cheng; Xin Zhang; Yu-Dong Zhang
Journal:  Comput Math Methods Med       Date:  2021-03-08       Impact factor: 2.238

8.  When less is more: regarding the use of chest X-ray instead of computed tomography in screening for pulmonary metastasis in postmolar gestational trophoblastic neoplasia.

Authors:  Antonio Braga; Kevin Meyer Elias; Neil Stuart Horowitz; Ross Stuart Berkowitz
Journal:  Br J Cancer       Date:  2021-01-08       Impact factor: 7.640

9.  Chest CT in patients with a moderate or high pretest probability of COVID-19 and negative swab.

Authors:  Caterina Giannitto; Federica Mrakic Sposta; Alessandro Repici; Giulia Vatteroni; Elena Casiraghi; Erminia Casari; Giorgio Maria Ferraroli; Alessandro Fugazza; Maria Teresa Sandri; Arturo Chiti; Balzarini Luca
Journal:  Radiol Med       Date:  2020-08-29       Impact factor: 6.313

View more
  14 in total

1.  Development and validation of bone-suppressed deep learning classification of COVID-19 presentation in chest radiographs.

Authors:  Ngo Fung Daniel Lam; Hongfei Sun; Liming Song; Dongrong Yang; Shaohua Zhi; Ge Ren; Pak Hei Chou; Shiu Bun Nelson Wan; Man Fung Esther Wong; King Kwong Chan; Hoi Ching Hailey Tsang; Feng-Ming Spring Kong; Yì Xiáng J Wáng; Jing Qin; Lawrence Wing Chi Chan; Michael Ying; Jing Cai
Journal:  Quant Imaging Med Surg       Date:  2022-07

2.  COVLIAS 2.0-cXAI: Cloud-Based Explainable Deep Learning System for COVID-19 Lesion Localization in Computed Tomography Scans.

Authors:  Jasjit S Suri; Sushant Agarwal; Gian Luca Chabert; Alessandro Carriero; Alessio Paschè; Pietro S C Danna; Luca Saba; Armin Mehmedović; Gavino Faa; Inder M Singh; Monika Turk; Paramjit S Chadha; Amer M Johri; Narendra N Khanna; Sophie Mavrogeni; John R Laird; Gyan Pareek; Martin Miner; David W Sobel; Antonella Balestrieri; Petros P Sfikakis; George Tsoulfas; Athanasios D Protogerou; Durga Prasanna Misra; Vikas Agarwal; George D Kitas; Jagjit S Teji; Mustafa Al-Maini; Surinder K Dhanjil; Andrew Nicolaides; Aditya Sharma; Vijay Rathore; Mostafa Fatemi; Azra Alizad; Pudukode R Krishnan; Ferenc Nagy; Zoltan Ruzsa; Mostafa M Fouda; Subbaram Naidu; Klaudija Viskovic; Mannudeep K Kalra
Journal:  Diagnostics (Basel)       Date:  2022-06-16

3.  MDFNet: an unsupervised lightweight network for ear print recognition.

Authors:  Oussama Aiadi; Belal Khaldi; Cheraa Saadeddine
Journal:  J Ambient Intell Humaniz Comput       Date:  2022-06-18

4.  A fast lightweight network for the discrimination of COVID-19 and pulmonary diseases.

Authors:  Oussama Aiadi; Belal Khaldi
Journal:  Biomed Signal Process Control       Date:  2022-06-21       Impact factor: 5.076

5.  Comparison and ensemble of 2D and 3D approaches for COVID-19 detection in CT images.

Authors:  Sara Atito Ali Ahmed; Mehmet Can Yavuz; Mehmet Umut Şen; Fatih Gülşen; Onur Tutar; Bora Korkmazer; Cesur Samancı; Sabri Şirolu; Rauf Hamid; Ali Ergun Eryürekli; Toghrul Mammadov; Berrin Yanikoglu
Journal:  Neurocomputing       Date:  2022-02-10       Impact factor: 5.779

Review 6.  Application of machine learning in CT images and X-rays of COVID-19 pneumonia.

Authors:  Fengjun Zhang
Journal:  Medicine (Baltimore)       Date:  2021-09-10       Impact factor: 1.817

7.  A novel fusion based convolutional neural network approach for classification of COVID-19 from chest X-ray images.

Authors:  Anubhav Sharma; Karamjeet Singh; Deepika Koundal
Journal:  Biomed Signal Process Control       Date:  2022-05-02       Impact factor: 5.076

8.  A complete framework for accurate recognition and prognosis of COVID-19 patients based on deep transfer learning and feature classification approach.

Authors:  Hossam Magdy Balaha; Eman M El-Gendy; Mahmoud M Saafan
Journal:  Artif Intell Rev       Date:  2022-01-29       Impact factor: 9.588

Review 9.  A COMPARATIVE STUDY OF X-RAY AND CT IMAGES IN COVID-19 DETECTION USING IMAGE PROCESSING AND DEEP LEARNING TECHNIQUES.

Authors:  H Mary Shyni; E Chitra
Journal:  Comput Methods Programs Biomed Update       Date:  2022-03-07

10.  Application of CycleGAN and transfer learning techniques for automated detection of COVID-19 using X-ray images.

Authors:  Ghazal Bargshady; Xujuan Zhou; Prabal Datta Barua; Raj Gururajan; Yuefeng Li; U Rajendra Acharya
Journal:  Pattern Recognit Lett       Date:  2021-12-03       Impact factor: 3.756

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.