Literature DB >> 35693545

FCF: Feature complement fusion network for detecting COVID-19 through CT scan images.

Shu Liang1, Rencan Nie1,2, Jinde Cao3,4, Xue Wang1, Gucheng Zhang1.   

Abstract

COVID-19 spreads and contracts people rapidly, to diagnose this disease accurately and timely is essential for quarantine and medical treatment. RT-PCR plays a crucial role in diagnosing the COVID-19, whereas computed tomography (CT) delivers a faster result when combining artificial assistance. Developing a Deep Learning classification model for detecting the COVID-19 through CT images is conducive to assisting doctors in consultation. We proposed a feature complement fusion network (FCF) for detecting COVID-19 through lung CT scan images. This framework can extract both local features and global features by CNN extractor and ViT extractor severally, which successfully complement the deficiency problem of the receptive field of the other. Due to the attention mechanism in our designed feature complement Transformer (FCT), extracted local and global feature embeddings achieve a better representation. We combined a supervised with a weakly supervised strategy to train our model, which can promote CNN to guide the VIT to converge faster. Finally, we got a 99.34% accuracy on our test set, which surpasses the current state-of-art popular classification model. Moreover, this proposed structure can easily extend to other classification tasks when changing other proper extractors.
© 2022 Elsevier B.V. All rights reserved.

Entities:  

Keywords:  COVID-19 detecting; Deep Learning; Feature complement fusion; Weakly supervised learning

Year:  2022        PMID: 35693545      PMCID: PMC9167685          DOI: 10.1016/j.asoc.2022.109111

Source DB:  PubMed          Journal:  Appl Soft Comput        ISSN: 1568-4946            Impact factor:   8.263


Introduction

Since Dec 2019, a coronavirus storm turned out in one night, and outbreaking spread all over the world rapidly. This type of coronavirus called Coronavirus Disease 2019 (COVID-19) became an ongoing epidemic engulfing the globe. COVID-19 was firstly reported in Wuhan, Hubei province of China [1]. Whereas, extensive transportation and population mobility were still operating during the Chinese Spring Festival in China, which aggravated the epidemic when people were unaware of its human-to-human transmission trait [2]. Some reports have indicated that this type of coronavirus has a low potential for sustained community transmission [3]. However, its spread speed is far more what people can imagine. According to the report, this COVID-19 epidemic follows an exponential growth [2], [3]. And up to November 2021, over 256,966,237 people have been detected infection, and deaths-cumulative have achieved 5,151,643 total [4]. This disease is highly contagious and may lead to acute respiratory distress or multiple organ failure in severe cases [5]. People infected with COVID-19 act fever and cough as the most common symptoms [6]. But these cannot examine the COVID-19 as the principal argument. Recent theoretical developments have revealed that the nearest clinical diagnosis regards the real-time reverse transcriptase-polymerase chain reaction (RT-PCR) as the golden standard to diagnose COVID-19 [7]. However, a scientifically synthetic diagnosis must combine Chest X-ray (CXR) or computed tomography (CT) result on patients. Both results show the same characteristic in COVID-19 symptoms, demonstrating bilateral peripheral consolidation in their lungs [8]. CXR and CT contribute to early diagnosing the viral disease, although nucleic acid detection with RT-PCR remains the standard Ref. [9]. Baseline CXR performed a sensitivity of 69% compared to 91% of RT-PCR [8]. Nevertheless, the CT scan receives a better sensitivity in contrast to the RT-PCR [9], [10]. Ground-glass opacity (GGO) and consolidation are two principal higher confidence characteristics for diagnosis, which display more distinctively on the CT scan than the CXR [8], [11]. Therefore CT scan displays a unicorn role in COVID-19 diagnose mission. With the development of computer science nowadays, applications based on the deep learning (DL) approach are extensively used around our life, especially in image processing for feature extraction purposes. Like object detection and identification based on local calculate service, or deploys feature extraction models on users and edges relying on cloud computing so-called federated learning [12]. And numerous industries began to add the DL technique to combine their original design. Convolutional neural network (CNN) especially shows its powerful ability in feature extraction work, which has dominated the DL field in recent decades. This type of network has even influenced the field of Clinical Medicine Science because its tremendous ability has been adopted widely in finding small details people overlooked [13]. Such as infection diagnosis: CNN assists the doctor in making the decision more precisely, which can focus on details people may neglect. It will deliver a prediction on the infection category or locate the position of the lesion [14], [15], [16]; lesion and organ segmentation: doctors once need segment the lesion and organ manually until CNN was involved in and support doctors to segment their interest part automatically for the succeeding diagnose [17], [18], [19]; multimodal medical image fusion: image imaging from different apparatus stores various information, whereas CNN fused images can reduce redundant data and preserve the essential data from both source images [20], [21], [22]. Nearly, Vision Transformer (ViT) appeared in the computer vision (CV) field, which migrates the algorithm used in natural language processing (NLP). This action impacts the status dominated by CNN and exerts a profound influence on CV researchers. In the subsequent time, ViT presented, researchers applied this structure to explore more facet applications. Like U-Transformer does a great framework on the task of complex organ segmentation [23]. And RTMIC can caption the CXR image automatically and generate a medical diagnosis [24]. A massive of methods of DL assist diagnosis occurred in this anti-COVID storm. Inspired by those works mentioned above, we further put forward our framework. In contrast to existing methods, the main contribution of our approach can be listed as follows: Different from CNN-based local feature extraction or ViT-based global feature extraction, our method employs two ones to extract local and global features of CT, respectively, and further integrated these features to give more effective discrimination for COVID-19. We construct a feature fusion block named FCT to fuse the local and global features. Due to the usage of the Transformer, our FCT can produce an embedding vector resulting from attention mechanism to present better feature representation. We designed a hybrid loss combined supervised learning with a weakly supervised learning strategy training our model efficiently to boost ViT structure converge faster in our proposed FCF. Our proposed FCF defeats other models on the benchmark and provides a reliable pre-diagnose to assist clinic surgeons doing the final diagnosis. The rest of the paper is organized as follows. In Section 2, we introduced the recent feature extraction works and feature fusion works. Section 3 included details of our algorithm methods. Dataset choice, training details, and extended experiments are arranged in Section 4. The conclusion is finally given in Section 5.

Related work

Feature extractors

CNN has been used widely in the DL involved task for feature extraction since VGG [25] opened up the era of the deep neural network. Subsequently, add identity mapping from the shallower layer into the deep layer first appeared in ResNet [26]. It was broadly adopted to maintain the feature in CNN encoders helpfully and overcome the gradient vanish in the training process [27]. Another framework to efficiently reuse features is DenseNet [28]. This novel approach leverage violent concatenate operation to link all features from the front layer to the current layer. It further reduces the parameters and accelerates the propagation efficiency in the deep neural network. Because of the development of the neural network, artificial intelligence for the COVID-19 detecting task joins following this popular stream as well. Yujin Oh et al. exploit the feature reuse by dense architecture to receive a clear edge result on COVID-19 CT lung segmentation [29]. People began to shift their attention to researching what information should be most careful when extracting information from images meets its bottleneck. For this reason, attention-based algorithms become embedded in blocks to lead crucial information into a better representation [30], [31]. Attention mechanisms combined with the residual block, which emerges henceforth, further activate to focus and preserve more on core information importance [32], [33]. But in recent years, a popular model stemmed from the natural language processing (NLP) domain called Transformer transplant into the computer vision domain successfully [34]. This kind of model named Vision Transformer (ViT) [35] is a fully-attention structure and inherent the conspicuous characteristic of BERT [36], which applies a token ahead of the embedding for final decision. However, the original ViT structure is hard to train economically without a large base dataset supported even if it surpasses the remarkable performance created by CNN [37]. And the converge situation of the Transformer model heavily depends on the training batch and multi-parallel GPUs [38]. Consequently, people began to add absolute position attribution on tokens, a different approach from the position embedding (PE). T2T-ViT compresses the tokens by re-structurization them like input images [37]. This approach provides a piece of new position information between tokens. Besides, the NesT creates tokens by block aggregation service to contribute sturdy position information. Besides, the NesT also emphasizes the importance of the absolute position by deploying block aggregation service on tokens [39]. In addition, tokens in the Swin-Transformer serve as different scales for acquiring extra position information [40]. These optimizations of ViT establish closer connections on position information, which made its long-range relationship characteristics involved orderly.

Feature fusion

Vary embeddings from the feature extractor represent different detailed information learned by the network. UNet fuse the same scale feature in the encoder and decoder network to reinforce the context feature [41]. Whereas feature pyramid network (FPN) fuse the different scale feature embeddings in the encoding process to acquire a better feature representation [42]. Another way is utilizing a fully-connected layer to process the concatenated embeddings, which also can conserve essential instances of them all [43]. In [44], Chao et al. encode multi-type data and concatenate the representation embeddings for synthesis prediction on ICU admission through the random forest. Anunay Gupta et al. proposed InstaCovNet obtains a more reliable accuracy performance by associating features from five different feature extractors together for the final decision [45].

Method

Motivation

Current works only applied CNN or ViT for feature extraction. But CNN is only good at capturing local features, which cannot extract global features causes of its small convolution kernels. In contrast, ViT benefits from its particular way to get embeddings from patches, which obtain a large receptive field but lack of detail information. So, we consider that integrating the feature from CNN and ViT can complement the feature deficiency of each other. The respective field of CNN and ViT were visualized in Fig. 1 (a) and Fig. 1(b) severely. According to the trait of their respective field, we define features from CNN as local features and features from ViT as global features. In Fig. 1(a), we have already cannot recognize what is meaning in the filter. Whereas we can clearly be aware of an outline with some details of a lung was shown in the filter as Fig. 1(b).
Fig. 1

Respective Field: (a) respective field of CNN, (b) respective field of ViT.

And we also consider that merging local feature embedding and global feature embedding can complement the data deficiency of each other. But operating a linear layer only to process the concatenated embeddings cannot emphasize the complement information specifically. We thus proposed a feature complement Transformer (FCT) to strengthen the expression of the concatenated feature embeddings. Due to our architecture is consists of a ViT, and the full-Transformer model needs a large dataset to train. Otherwise, it will exhibit an inferior performance compared to the identical size CNN model without pretraining [37]. And CT scan images are grayscale images, which cannot transfer learning on the three-channel input pretraining model. It also converges slowly when training on a single GPU without transfer learning. To resolve this problem mentioned above, we provided a weakly supervised module to guide the ViT extractor gaining better performance. Respective Field: (a) respective field of CNN, (b) respective field of ViT.

Network

According to our motivation in Section 3.1, we proposed a feature complement fusion network (FCF) contributing to providing a brand new view to get various aspect feature embeddings. And further combining the CNN and ViT hybrid embeddings integrated for prediction efficiently. The overview of our FCF structure is shown in Fig. 2. Pseudocodes of the whole FCF architecture are shown in 1. Specifically, we separate this model into three parts:
Fig. 2

The architecture of our proposed FCF-Net includes two feature extractors, a feature fusion block, and a weakly supervised module. A CT scan image passed two various feature extractors to generate 512-D local and global feature embeddings. Feature fusion block concatenates these embeddings and processes them within a feature complement Transformer (FCT), then makes the final prediction. The weakly supervised module first maps the various embedding to 2-D feature. These 2-D features will translate into probability distribution through a softmax layer. Finally, we calculate the KL divergence of these two distributions to optimize the ViT extractor in a weakly supervised way.

Feature extractor: this module consists of a CNN branch and a parallel ViT branch to extract CT images features into feature embeddings. Weakly supervised module: it produces a weak label to weakly supervise the ViT extractor, which accelerates its convergence rate meanwhile. Feature complement fusion block: features come from the feature extraction part gathered here, then make a complement fusion for the final decision. The architecture of our proposed FCF-Net includes two feature extractors, a feature fusion block, and a weakly supervised module. A CT scan image passed two various feature extractors to generate 512-D local and global feature embeddings. Feature fusion block concatenates these embeddings and processes them within a feature complement Transformer (FCT), then makes the final prediction. The weakly supervised module first maps the various embedding to 2-D feature. These 2-D features will translate into probability distribution through a softmax layer. Finally, we calculate the KL divergence of these two distributions to optimize the ViT extractor in a weakly supervised way.

Feature extractor

Features extracted from the feature extractor have to involve local information and global information from the object. We select CNN and ViT to extract local information and global information severally. CNN Extractor In our model, the CNN extractor is based on a backbone network, e.g., the ResNet [26] and DenseNet [28]. We will discuss how to select a specific one in the “Ablation Study” and “Compared to Classic SOTA” sections. Where to balance the compute efficiency and the performance, we finally adopt ResNet-50 as an extractor to extract local features with spatial information consisting of abundant texture and edge [37], [46] in the CNN branch. Generally, a ResNet-50 includes one 7 × 7 convolutional layer, 16 residual blocks, and a linear layer. Each residual block consists of 3 convolutional layers of kernel size 1 × 1, 3 × 3, and 1 × 1 respectively. A ReLU activate function is employed between every convolution layer. More details can see in [26]. Finally, we retain all convolutional blocks and rectify the output dimension of the final linear layer to generate a 512-D feature. ViT Extractor Concerning the limitation of the training condition mentioned in the “motivation” section, we utilized T2T-ViT [37] in the ViT extractor to extract global features by its long-range dependency. We make two modifications to the T2T-ViT so that it can fit COVID-19 classification tasks. Firstly, we truncated the inner Transformer layer into 5. Secondly, we adjust its hidden dimension to 512 formulating a 512-D feature directly. According to the feature extractor we chose above, a pair of 512-D feature embeddings generated from ResNet-50 and T2T-ViT-5, which represents local and global features severally.

Feature fusion block

We note that using concatenated feature embedding from CNN extractor and ViT extractor directly cannot reflect the importance proportion derived from different feature extractors. For this reason, we designed a feature complement Transformer (FCT) shown in Fig. 3 to conduct the merging features from the different feature extractors. FCT adopts the multi-head self-attention as the core mechanism, which gets the ability to concentrate on the essential value automatically. After that, we further introduced a single-layer feed-forward network (FFN) to reinforce the expression capability of the FCT structure.
Fig. 3

FCT structure.

Let denote the concatenated feature embedding from the CNN and ViT extractor: where is the length of the concatenated feature embedding. To meet the requirement of calculating the self-attention, we thus add an extra dimension to this embedding to get . This approach makes the embedding match the size of the FCT required only, which does not influence the data and its structure. To balance the computational efficiency and model performance, we finally utilize an 8-head self-attention structure, such that each instance project to embeddings , , and of each can represent as: where , , and denote the embedding weight matrix of each head. The process of calculating the attention of the whole feature embedding of each head is then given by: where is the embedding dimension of each head. The symbol () represents the dot product operation. For each of these, we refer to the final attention embedding as: where MLP represents the multilayer perceptron and GELU denotes the activation function. The output dimension of and is set to and correspondingly. Then restore its shape from to . Cause the attention mechanism effect, the processed feature embedding retains the crucial instance and attaches more importance to its staple feature from the different feature extractor. On the other side, the weakness instance, meanwhile, recedes its representation through the attention process. That means local feature and global feature reinforce their representation and integrate more efficiently. In addition, we add an FFN as the sub-layer to apply local and translationally equivariant that do not supply in the attention mechanisms. It further strengthens the representation power of the whole FCT structure. As other neural network classification tasks do, we adopt a linear classifier and a softmax layer to make the final prediction as: FCT structure.

Weakly supervised module

To overcome the convergence problem in the ViT structure, we designed this weakly supervised module. DL model for classification task final mapping the feature embedding into a probability distribution, no matter the CNN model or the ViT model. And considering the CNN model has tremendous generalization ability, we adopt the probability distribution created by the CNN extractor to be a weak label. We employ this weak label to inform the ViT where the probable location of proper distribution is, although it cannot provide a definite distribution of the ground truth. First, this module leverage the embedding features from two different extractors into 2-D features by a linear layer. Then we use a softmax layer to render it into a probability distribution of the final prediction. We use these distributions to calculate KL divergence to optimize the ViT extractor in training. It will detail more in the modified loss section.

Modified loss

We utilize the categorical cross-entropy loss as the majority loss to supervise this classification model. where  is the actual probability of the image,  denotes the predict label of the model. Consider our model incorporated with ViT structure for adding global features. And Transformer-based model gets a worse performance when it lack of large dataset to do pretraining on it, in contrast to the same size CNN model counterpart [37], [47]. To address this problem, we created a weakly supervised module to accelerate the rate of convergence of ViT by a weakly supervised learning manner. Due to these two distributions from various feature extractors needing to describe the same prediction result, we expected the prediction distribution from these feature embeddings to be similar. We consequently introduced a KL divergence as a regularization term loss to minimize the distance of two distributions: where  here denotes each input from the dataset , means the distribution of generated from the CNN feature embedding. can be obtained in the same way by the ViT extractor. We set as the pseudo label, let learn information from the pseudo label to optimize the ViT extractor. That means the ViT extractor can learn the distribution generated from the CNN extractor to accelerate its convergence rate and train the ViT extractor more efficiently. The total loss can be written as follow: where represents an impact factor to restrain how strength this term intervenes in the ViT extractor optimization process. will be discussed in the following parameter analysis section.

Experiment

Implementation details

Dataset. The dataset we used is a subset of COVID-CTset, which was introduced in a COVID-19 pneumonia detecting framework by Mohammad Rahimzadeh [48]. COVID-CTset was gathered from Negin medical center in Iran, they adopt a SOMATOM Scope model and syngo CT VC30-easyIQ software version for capturing and visualizing the lung HRCT radiology images from the patients. It uses 16-bit rather than 8-bit grayscale images to reserve more detail for experts to diagnose low-level inflation precisely. These kinds of images in this dataset perfectly evade the data leakage problem by compressing the image into other formats mentioned in [49]. CT images are not in color, and these raw data is only one channel instead of RGB (R=G=B) [49], which is more scientific for the later training process. This subset contains 4562 images which controlled the positive and negative samples as the same proportion strictly. And then we split the selected images into train-set, val-set and test-set as a proportion of 8:1:1. We show the dataset configuration in Fig. 4 and sample details in Fig. 5.
Fig. 4

Dataset Configuration.

Fig. 5

Samples in COVID-CTset: (a) COVID-19 samples, where red arrows denote the infection area. (b) Normal samples.

Training details. The implementation of the whole FCF architecture software is based on the Pytorch [50] framework. The hardware information to train this model is on an i9-10900F CPU and a single RTX 3090 GPU. Our proposed work is an end-to-end model, which represents this model requires a CT image only for the final result. We train this network end-to-end for 200 epochs using Adam [51] optimizer with a constant learning rate of 7e-5 through our proposed loss. The parameters in the Adam optimizer are setting as 0.9, 0.999. Dataset Configuration. Samples in COVID-CTset: (a) COVID-19 samples, where red arrows denote the infection area. (b) Normal samples. Assessment criteria. And to better critic the performance of our proposed model, we applied four evaluation metrics to judge our model synthetically. where True Positive (TP): The model forecasts the COVID-19 category to be the COVID-19 class. False Positive (FP): The model predicts the initially Normal class into the COVID-19 class. True Negative (TN): The model forecasts the Normal category to be the Normal class. False Negative (FN): The model predicts the COVID-19 class into the Normal class.

Parameter analysis

We introduced a hyperparameter to describe the impact force on the regularization term in Eq. (11). This parameter constrains how similar these distributions are, generated from the different extractors. We experiment with seven different values of this hyperparameter from 0.4 to 1.2 to analyze the performance of the COVID-19 detection model. To guarantee fairness in the experiment, we initial the training model at the same weight. And we introduced four metric methods from Eq. (15) to judge the result synthetically. These four metric methods show different variations when changing the parameter . We can see the best accuracy shows up s.t. = 0.6 in Table 1. It indicates that this model can make the minimum error decision in the prediction process. The highest recall emerged when = 1.0, which represents the model acquired an excellent capability to cover the fully COVID-19 samples in the test set. Besides, the precision reaches the apogee when = 0.6. It demonstrates our model gets the best ability to predict the COVID-19 class using this hyperparameter. However, greedy much on the recall or precision of the model is not a wise choice. The comprehensive critical ability of the focal infection prediction model cannot be neglected. Thus we compared their F1-Score, which employs an average judgment on recall and precision, to determine the best accuracy and F1-Score model. We finally choose = 0.6 to be the best optimization hyperparameter.
Table 1

Parameter Analysis.

Para.ValueAccuracyRecallPrecisionF1-Score
λ0.497.8198.2597.3997.82
0.598.6898.2599.1298.68
0.699.3499.5699.1399.34
0.797.8196.9398.6697.79
0.897.1596.4997.7897.13
1.098.90100.0097.8598.92
1.298.4697.8199.1198.45
Parameter Analysis.

Ablation study

We design various ablation experiments to illustrate the effectiveness of our proposed methods in this brand-new architecture. We show the ablation result of the whole architecture used ResNet-50 to be the CNN extractor and T2T-ViT-5 as the ViT extractor in Table 2. All these experiments are based on the same initial weight to ensure fairness. We first focus on how the KL divergence in the loss function affects the performance of the whole architecture. Then, we remove the KL divergence in the loss function and add our FCT structure to test the influence of this proposed method. After that, we merged them into the model to see its bonus performance.
Table 2

Ablation on FCF-Res model.

FCTKLAccuracyRecallPrecisionF1-Score
97.8197.3798.2397.80
98.4698.6898.2598.47
98.4998.6898.2598.47
99.3499.5699.1399.34
KL divergence. We introduced KL divergence in the loss function to assist the ViT extractor in learning from the CNN extractor. This approach aims to train the model more efficiently and overcome the low convergence defect of the Transformer-based model. It raises the accuracy performance in the final prediction as well. Although ablation on KL divergence only model gets the same performance as the FCT only model. As shown in Fig. 6, the model adding KL divergence to the loss function achieves a higher speed in the convergence course.
Fig. 6

The variation of accuracy in the training process: (a) FCF-Res model, (b) FCF-Dense model.

Ablation on FCF-Res model. Ablation on FCF-Dense model. FCT. Next, we dissect the result in Table 2 and note that the whole model performed well when adding our elegant fusion strategy — FCT into it. Given result demonstrates that using FCT rather than mere concatenated feature embeddings can make a 0.25% rise in prediction accuracy. It also developed the recall indices, which represent the whole model receive a strong ability in detecting disease through the FCT. We do the same ablation experiment when replacing the ResNet-50 with DenseNet-121 in CNN branch to prove the validity of our framework. The result in Table 3 corroborates the effectiveness of our thought and the correctness of our framework as well.
Table 3

Ablation on FCF-Dense model.

FCTKLAccuracyRecallPrecisionF1-Score
98.4699.1297.8498.47
98.9099.1298.6998.91
98.6899.5697.8598.70
99.3499.1299.5699.34
The variation of accuracy in the training process: (a) FCF-Res model, (b) FCF-Dense model.

Compared to classic SOTA

We do experiments to compare our network with the state-of-the-art classic classification algorithms, including CNN models, ViT models, and feature fusion models in Table 4.
Table 4

Compared to Classic SOTA.

Model typeModelParams. (M)FLOPs (G)AccuracyRecallPrecisionF1-Score
CNNResNet 5023.510.697.8197.3798.2397.80
ResNet 10142.520.298.0398.2597.8298.03
ResNet 15258.130.098.6899.5697.8498.70
DenseNet 1216.97.298.4699.1297.8498.47
DenseNet 16126.520.098.4699.5697.4298.48
DenseNet 16912.58.698.9099.1298.6998.91
DenseNet 20118.111.298.6899.5697.8498.70

ViTViT-b15.88.281.3688.1677.6082.55
T2T-ViT-510.96.094.7492.1197.2294.59

Feature FusionFPN(ResNet 50)26.317.898.0397.8198.2498.02
UNet17.280.099.3499.56100.099.34

Feature Complement FusionFCF-Res(ours)40.016.699.3499.5699.1399.34
FCF-Dense(ours)24.213.299.3499.1299.5699.34
FCF reaches 99.34 accuracy, which is 0.44 higher when compared to CNN models. This performance can defeat any CNN model, but it receives more parameters and FLOPs than ResNet-50 and DenseNet all series. As we declared in the related work, a fully-Transformer structure will get a poor performance if it encounters a data scarcity condition or training on a single GPU. Hence, we only exhibit the result of base ViT and the T2T-ViT-5 we used in our framework. As for FPN and UNet, they both gathered features to enhance the representation for the final prediction. FPN concatenates every embedding from different fusion layers in the pyramid network. And every sub-layer under the top layer applied a feature enhancement by adding an up sample feature of its up-neighbor layer. This process increases 40 FLOPs than normal ResNet-50, although it gets a better accuracy performance when compared to ResNet-50. UNet attains the same high accuracy with 99.34, according to its characteristic U-shape structure. This architecture applied huge feature maps in the skip connection process to do feature enhancement roughly. However, the FLOPs run up rapidly along with that process corresponding. A four times increase in FLOPs to get an extra 0.22 F1-Score is not a wise choice. In contrast, our provided framework makes a good trade-off on prediction performance and computation efficiency. We sacrifice the training parameters and bring our model size into the middle-level to trade fewer FLOPs for refining our model. The confusion matrix we post in Fig. 7 shows the detailed performance of each model we applied in our feature complement fusion framework. We compare each class that was predicted, including every engaged sample. Although T2T-ViT-5 performs inferior to ResNet-50 and DenseNet-121, it provides useful property features fed into FCT and promotes prediction performance. Compare the data shown in Fig. 7 (a) and Fig. 7 (d), we can conclude that global features generated from T2T-ViT-5 compensate the local features extracted by the ResNet-50. The same conclusion can be easily obtained when comparing the Fig. 7 (b) and Fig. 7 (e). The AUC of ROC curve results in Fig. 8 and Fig. 9 also shores up the effectiveness of our proposed model exactly.
Fig. 7

Confusion matrix of FCF and their source feature extractor.

Fig. 8

ROC of FCF and their source feature extractor.

Fig. 9

ROC of FCF and other feature fusion methods.

Compared to Classic SOTA. Confusion matrix of FCF and their source feature extractor. ROC of FCF and their source feature extractor. ROC of FCF and other feature fusion methods. Grad-CAM on comparison methods. Accuracy performance of COVID-19 SOTA methods on COVID-CTset [48], iCTCF [52] and MosMedData [53].

Compared to COVID-19 SOTA

We also compared our FCF-Res with other COVID-19 state-of-art works on three public datasets to verify its generalization ability and effectiveness. As shown in Table 5, InstaCovNet-19 and our FCF-Res emerge with the same accuracy of 99.34% on the COVID-CTset, but our calculation efficiency is more economical. Whereas InstaCovNet-19 integrated five baseline models to get the final prediction. Due to InstaCovNet-19 integrating five baseline models to get the final prediction, it boosts parameters and FLOPs dramatically, although it achieves the best performance on iCTCF [52]. It boosts parameters and FLOPs increasing to a dramatic extent, although it achieves the best performance. Four times of resource usage but only increasing the accuracy by 1.12% is not a cost-effective strategy. COVID-Net applied the minimum parameters, which perform well on iCTCF while lacking generalization ability performing worse on other datasets. Even on the MosMedData [53], our model shows the best accuracy performance and verified its robustness and generalization ability.
Table 5

Accuracy performance of COVID-19 SOTA methods on COVID-CTset [48], iCTCF [52] and MosMedData [53].

MethodParams. (M)FLOPs (G)COVID-CTset[48]iCTCF [52]MosMedData [53]
ADLN [48]26.217.896.0597.5672.40
InstaCovNet-19 [45]165.869.299.3499.3178.73
COVID-Net CT [54]15.20.978.8297.0066.86
FCF-Res(ours)40.016.699.3498.1984.50
We also handle Grad-CAM [55] to visualize the result to illustrate where the gradient actually focused in Fig. 10. ADLN shows a similar result on iCTCF (Fig. 10 (c)) as ours (Fig. 10 (f)) but appears weak gradient on MosMedData. Although InstaCovNet-19 achieves the best accuracy, it cannot locate the gradient of lesion areas in visualization results. The gradient of COVID-Net CT cannot focus on lesion areas in all three datasets, which also explains its poor generalization ability. The gradient was relatively dispersed in the T2T-ViT-5 or made misregistration of lesion areas in ResNet-50. Ours FCF integrated the advantage characteristic of ResNet-50 and T2T-ViT-5. It not only possesses broader receptive fields to guarantee a wide range of gradients but also preserves effective gradients on the lesion areas precisely.
Fig. 10

Grad-CAM on comparison methods.

Conclusion

In this paper, we introduce a detailed study of feature complement fusion network (FCF) to detect COVID-19 through CT images. This structure involves the advantage of the receptive field in both CNN and ViT models to extract complement features. And a weakly supervised module successfully assists the ViT structure to converge better. We further provided an FCT architecture to merge the characteristics from feature extractors economically. Extensive ablation experiments demonstrate that our proposed method specifically worked in the weakly supervised module and feature fusion block. Moreover, these algorithms perform well in the whole construction. It emphasizes the priority in different features and enables them fused more efficiently. Although we alleviate the converge slow question caused by ViT in our architecture, our model is training unstable cause of its inner layer normalization structure. It also needs a long time to train for convergence on a large-scale dataset compared to the pure CNN model. We are going to overcome the problems mentioned above in our future works.

CRediT authorship contribution statement

Shu Liang: Conceptualization, Software, Writing – original draft and editing. Rencan Nie: Supervision, Writing – reviewing. Jinde Cao: Supervision. Xue Wang: Software. Gucheng Zhang: Validation.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  30 in total

1.  Dynamic-Fusion-Based Federated Learning for COVID-19 Detection.

Authors:  Weishan Zhang; Tao Zhou; Qinghua Lu; Xiao Wang; Chunsheng Zhu; Haoyun Sun; Zhipeng Wang; Sin Kit Lo; Fei-Yue Wang
Journal:  IEEE Internet Things J       Date:  2021-02-04       Impact factor: 10.238

2.  Relational Modeling for Robust and Efficient Pulmonary Lobe Segmentation in CT Scans.

Authors:  Weiyi Xie; Colin Jacobs; Jean-Paul Charbonnier; Bram van Ginneken
Journal:  IEEE Trans Med Imaging       Date:  2020-08       Impact factor: 10.048

3.  A Weakly-Supervised Framework for COVID-19 Classification and Lesion Localization From Chest CT.

Authors:  Xinggang Wang; Xianbo Deng; Qing Fu; Qiang Zhou; Jiapei Feng; Hui Ma; Wenyu Liu; Chuansheng Zheng
Journal:  IEEE Trans Med Imaging       Date:  2020-08       Impact factor: 10.048

4.  Checklist for responsible deep learning modeling of medical images based on COVID-19 detection studies.

Authors:  Weronika Hryniewska; Przemysaw Bombiski; Patryk Szatkowski; Paulina Tomaszewska; Artur Przelaskowski; Przemysaw Biecek
Journal:  Pattern Recognit       Date:  2021-05-21       Impact factor: 7.740

5.  Correlation of Chest CT and RT-PCR Testing for Coronavirus Disease 2019 (COVID-19) in China: A Report of 1014 Cases.

Authors:  Tao Ai; Zhenlu Yang; Hongyan Hou; Chenao Zhan; Chong Chen; Wenzhi Lv; Qian Tao; Ziyong Sun; Liming Xia
Journal:  Radiology       Date:  2020-02-26       Impact factor: 11.105

6.  CT Imaging Features of 2019 Novel Coronavirus (2019-nCoV).

Authors:  Michael Chung; Adam Bernheim; Xueyan Mei; Ning Zhang; Mingqian Huang; Xianjun Zeng; Jiufa Cui; Wenjian Xu; Yang Yang; Zahi A Fayad; Adam Jacobi; Kunwei Li; Shaolin Li; Hong Shan
Journal:  Radiology       Date:  2020-02-04       Impact factor: 11.105

7.  A new approach for computer-aided detection of coronavirus (COVID-19) from CT and X-ray images using machine learning methods.

Authors:  Ahmet Saygılı
Journal:  Appl Soft Comput       Date:  2021-03-17       Impact factor: 6.725

8.  Frequency and Distribution of Chest Radiographic Findings in Patients Positive for COVID-19.

Authors:  Ho Yuen Frank Wong; Hiu Yin Sonia Lam; Ambrose Ho-Tung Fong; Siu Ting Leung; Thomas Wing-Yan Chin; Christine Shing Yen Lo; Macy Mei-Sze Lui; Jonan Chun Yin Lee; Keith Wan-Hang Chiu; Tom Wai-Hin Chung; Elaine Yuen Phin Lee; Eric Yuk Fai Wan; Ivan Fan Ngai Hung; Tina Poy Wing Lam; Michael D Kuo; Ming-Yen Ng
Journal:  Radiology       Date:  2020-03-27       Impact factor: 11.105

9.  Chest CT for Typical Coronavirus Disease 2019 (COVID-19) Pneumonia: Relationship to Negative RT-PCR Testing.

Authors:  Xingzhi Xie; Zheng Zhong; Wei Zhao; Chao Zheng; Fei Wang; Jun Liu
Journal:  Radiology       Date:  2020-02-12       Impact factor: 11.105

10.  Integrative analysis for COVID-19 patient outcome prediction.

Authors:  Hanqing Chao; Xi Fang; Jiajin Zhang; Fatemeh Homayounieh; Chiara D Arru; Subba R Digumarthy; Rosa Babaei; Hadi K Mobin; Iman Mohseni; Luca Saba; Alessandro Carriero; Zeno Falaschi; Alessio Pasche; Ge Wang; Mannudeep K Kalra; Pingkun Yan
Journal:  Med Image Anal       Date:  2020-10-13       Impact factor: 8.545

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.