Literature DB >> 33828610

Diagnosing Covid-19 chest x-rays with a lightweight truncated DenseNet with partial layer freezing and feature fusion.

Francis Jesmar P Montalbo1.   

Abstract

Due to the unforeseen turn of events, our world has undergone another global pandemic from a highly contagious novel coronavirus named COVID-19. The novel virus inflames the lungs similarly to Pneumonia, making it challenging to diagnose. Currently, the common standard to diagnose the virus's presence from an individual is using a molecular real-time Reverse-Transcription Polymerase Chain Reaction (rRT-PCR) test from fluids acquired through nasal swabs. Such a test is difficult to acquire in most underdeveloped countries with a few experts that can perform the test. As a substitute, the widely available Chest X-Ray (CXR) became an alternative to rule out the virus. However, such a method does not come easy as the virus still possesses unknown characteristics that even experienced radiologists and other medical experts find difficult to diagnose through CXRs. Several studies have recently used computer-aided methods to automate and improve such diagnosis of CXRs through Artificial Intelligence (AI) based on computer vision and Deep Convolutional Neural Networks (DCNN), which some require heavy processing costs and other tedious methods to produce. Therefore, this work proposed the Fused-DenseNet-Tiny, a lightweight DCNN model based on a densely connected neural network (DenseNet) truncated and concatenated. The model trained to learn CXR features based on transfer learning, partial layer freezing, and feature fusion. Upon evaluation, the proposed model achieved a remarkable 97.99 % accuracy, with only 1.2 million parameters and a shorter end-to-end structure. It has also shown better performance than some existing studies and other massive state-of-the-art models that diagnosed COVID-19 from CXRs.
© 2021 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  AP, Average Pooling; AUC, Area Under the Curve; BN, Batch Normalization; BS, Batch Size; CAD, Computer-Aided Diagnosis; CCE, Categorical Cross-Entropy; CNN, Convolutional Neural Networks; CT, Computer Tomography; CV, Computer Vision; CXR, Chest X-Rays; Chest x-rays; Computer-aided diagnosis; Covid-19; DCNN, Deep Convolutional Neural Networks; DL, Deep Learning; DR, Dropout Rate; Deep learning; Densely connected neural networks; GAP, Global Average Pooling; GRAD-CAM, Gradient-Weighted Class Activation Maps; JPG, Joint Photographic Group; LR, Learning Rate; MP, Max-Pooling; P-R, Precision-Recall; PEPX, Projection-Expansion-Projection-Extension; ROC, Receiver Operating Characteristic; ReLU, Rectified Linear Unit; SGD, Stochastic Gradient Descent; WHO, World Health Organization; rRT-PCR, real-time Reverse-Transcription Polymerase Chain Reaction

Year:  2021        PMID: 33828610      PMCID: PMC8015405          DOI: 10.1016/j.bspc.2021.102583

Source DB:  PubMed          Journal:  Biomed Signal Process Control        ISSN: 1746-8094            Impact factor:   3.880


Introduction

In an unexpected turn of events, our world is once again experiencing another historical struggle from a novel coronavirus, namely SARS-CoV-2. As stated in a study, the origin of the virus started in Wuhan, China, around December 2019. Shortly after, the coronavirus became known as COVID-19 and had spread across nearby countries within a short period, and soon after, around March 2020, the World Health Organization (WHO) declared a global pandemic [1]. According to recent studies, COVID-19 can cause inflammation to the lungs that radiate common symptoms of high fever, dry cough, and fatigue. In some cases, patients may also experience the loss of taste, sore throat, and even skin rashes. However, due to the virus penetrating the lungs more than the other organs, people with weaker immune systems suffer from severe complications like shortness of breath, chest aches, and decreased mobility [2]. With the use of a diagnostic procedure, the real-time Reverse Transcription Polymerase Chain Reaction (rRT-PRC) can detect signs of COVID-19 from specimens acquired from nasopharyngeal or oropharyngeal swabs [3]. However, to perform such a test, the process requires a well-trained medical practitioner equipped with specialized equipment. Even today, some countries still lack the said expertise and diagnostic equipment, specifically in underdeveloped countries, and it continues to worsen as the spread of the virus progresses [4]. In such cases, due to the lack of the said testing equipment and personnel, like most common lung diseases, potential COVID-19 patients can also receive screenings through a lesser alternative like Chest X-Rays (CXR), compared to an rRT-PRC or a Computer Tomography (CT) scan. In some well-developed countries, having experts and the said technologies helps them cope with diagnosing COVID-19 in various methods. However, even with the X-rays' availability, the results still require high proficiency and experience to identify signs of COVID-19 infections accurately. Even for some expert radiologists, producing an accurate diagnosis specifically for COVID-19, is still complicated. The lack of such tends to cause slow, inaccurate, and higher examination costs [5,6]. Recently, technology progressed at an unimaginable rate. With the help of Computer-Aided Diagnosis (CAD), Deep Learning (DL), and Computer Vision (CV) methods, several studies produced ways to reduce such difficulty in the medical field. A study initiated by Wang et al. [7] proposed a solution using the mentioned methods to develop a Convolutional Neural Network (CNN) called the COVID-Net, explicitly designed to identify COVID-19 infected CXRs automatically from a curated dataset of various CXRs. Their proposed COVID-Net consisted of a lightweight residual pattern called Projection → Expansion → Projection → extension (PEPX). Their PEPX module utilizes 1 × 1 convolutional filters to address the image's lower dimensions at the first phase. With their added expansion method, the images expand differently from their original input form. A 3 × 3 depth-wise convolution contributes further by extracting additional image characteristics with less learning complexity and computational strain. The following projection phase then reverts the enlarged image to its previous dimension and expands the entire feature set's depth. With a test data of 300 images composed of three CXR classes, namely, Normal, COVID-19, and Pneumonia, their proposed COVID-Net achieved an overall accuracy of 93.3 % in detecting COVID-19 from CXRs. Their study also yielded better performance than VGG19 with 83.0 % and ResNet50 with 90.6 %. Unlike COVID-Net, Al-Falluji resolved the same problem using a more extensive pre-trained Deep Convolutional Neural Network (DCNN) model, ResNet-18 [8]. Their proposed model had modifications to suit the given problem by adding a noise filter, a Global Average Pooling (GAP) layer, and compression layers. For an improved learning process, they used transfer learning to acquire readily available features to improve the model's image recognition and train it in a short period of only 30 epochs [9]. Such a method also improved their model's overall performance compared to the standard handcrafted feature extraction approach. Upon evaluation, their proposed modified ResNet-18 attained an accuracy of 96.73 % from the classification of non-infected, COVID-19, and Pneumonia CXRs. Rather than relying on a single DCNN for the task of detecting COVID-19 from CXRs, Chowdhury et al. had the initiative to use an ensemble method [10]. Their study chose a recent DCNN model named EfficientNet as their base model. With transfer learning and fine-tuning, they managed to train several EfficientNet models to recognize several CXRs, including COVID-19 cases. Their trained EfficientNet models then classified samples of CXRs and had their prediction scores averaged as an ensemble. As stated in their results, their proposed ensembled EfficinetNet with a coefficient of three achieved 97 % accuracy with a low overhead demand than most state-of-the-art DCNNs that performed the same task. In another study by Singh et al. [11], they proposed a modified structure of the Xception model comprised of six convolution layers and twelve depth-wise convolutions. Their proposed model trained for 100 epochs using specialized hardware, augmentation techniques, and hyper-parameter tuning. Their combined methods achieved a 95.80 % accuracy that outperformed other studies like the DarkCovidNet [12] with 87.02 %, DeTraC-ResNet18 [13] with 95.12 %, and a Hierarchical EfficientNetB3 [14] with 93.51 %. Table 1 presents a summary of the discussed recent studies that performed the detection of COVID-19 infected CXRs. As discussed, the recent studies yielded exceptional performances towards the diagnosis of COVID-19 from CXRs using various methods and techniques. However, other methods can still induce improvements to reduce the computational cost and optimization efforts further while attaining better accuracy. Therefore, this work proposes a lightweight model using various techniques like network truncation, partial layer freezing, and feature fusion of a pre-trained DCNN model to diagnose COVID-19 infected CXRs.
Table 1

Summary of studies that diagnosed COVID-19 chest x-rays with DCNNs.

ModelAccuracy (%)ClassesType
COVID-Net [7]93.30Normal, COVID-19, PneumoniaCXR
Modified ResNet-18 [8]96.37Normal, COVID-19, PneumoniaCXR
ECOVNet-EfficientNetB3 base [10]97.00Normal, COVID-19, PneumoniaCXR
Modified Xception [11]95.70Normal, COVID-19, PneumoniaCXR
DarkCovidNet [12]87.02Normal, COVID-19, PneumoniaCXR
DeTraC-ResNet18 [13]95.12Normal, COVID-19, SARSCXR
Hierarchical EfficientNetB3 [14]93.51Normal, COVID-19, PneumoniaCXR
Summary of studies that diagnosed COVID-19 chest x-rays with DCNNs. Primarily, this work strongly contributes based on the following: This work truncated a pre-trained DCNN, the DenseNet121 model. The said method further reduced its network size and complexity without sacrificing a significant fraction of its performance. With that said, the truncated DenseNet attained a faster training time due to a shorter end-to-end structure while maintaining a rich extraction of relevant features. Compared to other studies, this work fused a mirror image of a truncated DenseNet model, where the other half had its entire network fully re-trained from a curated CXR dataset with COVID-19 cases and ImageNet. On the other hand, the other half had its upper layers partially frozen to generate a different feature batch. Both models' distinct features are then fused right before the network's end, handled by a proposed set of layers equipped with regularization layers to prevent overfitting problems. This method provided a broad spectrum of diverse features compared to a traditionally trained DCNN model. Unlike most existing works that performed the detection of COVID-19 from CXRs, this proposed method achieved remarkable results. It even outperformed other state-of-the-art models and studies without the need for heavy optimization, data augmentation, specialized hardware, long training epochs, ensemble methods, and an expensive computing cost. With that said, the proposed method can become easily deployable, reproducible, upgradeable, and most importantly, usable in future cases.

Materials and methods

Dataset collection and preparation

Due to the difficulty of acquiring CXRs, more or so samples with COVID-19, this work relied on a readily available curated dataset from Sait et al. [15]. Fig. 1 illustrates samples from the sourced dataset containing 3270 Normal (a), 1281 COVID-19 infected (b), and 4657 Pneumonia (c) CXR images.
Fig. 1

Sample images of normal (a), covid-19 infected (b), and (c) pneumonia chest x-rays.

Sample images of normal (a), covid-19 infected (b), and (c) pneumonia chest x-rays. The images had no fixed dimensions as all came from various reliable sources. With that said, this work had all images resized and normalized into a 299 × 299 dimension, maintained to its Joint Photographic Group (JPG) format to prevent large consumptions of memory and to add better stability during both training and validation phases [16]. In Table 2 , the entire collection of 9208 CXR images had proper distributions to train and validate the proposed model, in which 80 % of the entire dataset became the train data and the remaining 20 % as the validation data selected stochastically to prevent the case of bias [17]. It is worth mentioning that this work relied only based on the presented number of CXR samples to showcase the proposed model's performance towards an unbalanced dataset without relying on data augmentation.
Table 2

Specification of the curated dataset, with Normal, COVID-19, and Pneumonia chest x-rays.

Class labelTrain (80 %)Validation (20 %)Total (100 %)
Normal26166543270
COVID-1910252561281
Pneumonia (Bacterial and Viral)37269314657
Total736718419208
Specification of the curated dataset, with Normal, COVID-19, and Pneumonia chest x-rays.

Proposed model and construction

This section focuses on the proposed model's developmental process for the automated diagnosis of COVID-19 and the other mentioned CXRs. The following includes the DenseNet model's background that served as the base structure and concept, added with this work's proposed method that decreased the model's parameter size that still sufficiently generated a robust feature pool.

DenseNet

For the proposed model, this work selected to use a densely connected neural network, the DenseNet model, illustrated in Fig. 2 . The purpose of having a DenseNet model delivers a flawless propagation of features throughout the entire network without the saturation of performance, even with a more extensive depth. DenseNet also solves the parameter inflation through concatenation rather than the addition of layers [18].
Fig. 2

A simplified visual concept of the DenseNet model [18].

A simplified visual concept of the DenseNet model [18]. Fig. 3 illustrates an in-depth view of a 5-layer DenseNet with a k = 4 number of dense blocks. As presented, the DenseNet model begins with an input block (layer 1) composed of a 7 × 7 Convolutional (Conv) layer → Batch Normalization (BN) → Rectified Linear Unit Activation (ReLU) → 3 × 3 Max-Pooling (MP) layer. Subsequently, a set of dense blocks follows with a BN → ReLU → 1 × 1 Conv layer pattern followed by another BN → ReLU → 3 × 3 Conv. Unlike residual networks and other deep models consisting of massive parameters via feature summation, the DenseNet model uses a dense block with a k growth rate concatenated to all network layers. This technique became an effective transmission of feature inputs from the preceding layers towards the other succeeding layers end-to-end [18]. Such a design generates a rich gradient quality even at greater depths while sustaining a low parameter count, making it an ideal choice for this task.
Fig. 3

A DenseNet model that presents the internal specifications of its dense block and transition layer.

A DenseNet model that presents the internal specifications of its dense block and transition layer. Like most DCNN models, the DenseNet model requires a downsampling layer to prevent exhaustion of resources during feature extraction [18], where it utilizes a transition layer that downsamples the feature maps using a 1 × 1 Conv and a 2 × 2 Average Pooling (AP) with strides of 1 and 2, respectively.

Truncation method

Though the DenseNet model already has significantly fewer parameters than most DCNNs, the proposed method aims to reduce the parameters further without affecting too much of its performance, considering that even the smallest member of the DenseNet family, the DenseNet121, still has about 7 million parameters [18]. Since the goal of the original DenseNet revolves around massive datasets like the ImageNet with more than 14 million images and 1000 classes, training and reproducing this model can become tedious due to the lack of adequate computing resources. Also, with the limited dataset for this task, using the entire model's structure only adds complexity and consumes immense resources. Therefore, through a proposed truncation of its entire structure, most of its layers were removed, reducing the number of parameters and further shortened the end-to-end flow of features. Fig. 4 presents the proposed truncated DenseNet-Tiny with only six dense blocks followed by a transition layer connected to another set of three dense blocks left. The dedicated output layers will then connect when the model had fused. The proposed design significantly decreased the parameter size and feature depth of the DenseNet121 model by 93 %. From the initial 8 million parameters, DenseNet121, the truncated DenseNet-Tiny had reduced its parameters to only about half a million.
Fig. 4

The structure of the proposed truncated DenseNet-Tiny.

The structure of the proposed truncated DenseNet-Tiny.

Model concatenation and feature fusion

Due to the reduced network size of the truncated DenseNet-Tiny, it resulted in a smaller parameter size. In this case, the relatively small parameter size emanated consequences that affected the DenseNet-Tiny's performance in recognizing the three classes of CXRs effectively [19]. However, putting additional depth of layers can only make the truncation method pointless and distort the extraction of features. Therefore, an additional proposed method referred to as model concatenation [20] and feature fusion [21] provided insights that solved this problem. In Fig. 5 , a mirror image of the DenseNet-Tiny with the exact specifications combined to form the Fused-DenseNet-Tiny pipeline. Through model concatenation and feature fusion, the Fused-DenseNet-Tiny became wider instead of longer, maintaining the desired rapid end-to-end extraction of features during training and validation. Also, to add better support in handling the robust features produced by the fused model, this work added a new set of layers consisting of a GAP [22], Dense with 512 units with a ReLU activation, and Dropout [23] connected to another Dense layer with three units activated by a Softmax classifier [24]. The added layers aim to induce an improved performance and regularization that prevents overfitting issues [25], potentially allowing the model to perform better even with the provided unseen validation dataset and other future real-world CXR images.
Fig. 5

The blueprint of the proposed Fused-DenseNet-Tiny.

The blueprint of the proposed Fused-DenseNet-Tiny. Such a practical and less tedious solution provides an advantage of improving the feature production without designing or using another model that can re-initiate complexity in the network.

Transfer learning, fine-tuning, and partial layer freezing

Due to the scarcity of the desired CXR images for training, this work employed both transfer learning [9] and fine-tuning [26] to supply additional features for both DenseNet-Tinys. However, transferring the same set of features from ImageNet for both pipelines may only produce redundant features. Therefore, this work employed a fine-tuning technique that partially froze specific layers in one of the mirrored DenseNet-Tiny's network to produce independent features [27]. Fig. 6 illustrates that after the transfer learning process, instead of allowing both DenseNet-Tinys to re-train and extract features from the CXR dataset entirely, the other half had its extraction layers frozen. Unlike other studies, this work had a mirrored version fused, dubbed as the DenseNet-Tiny-B, where all its layers thawed and re-trained simultaneously to received newly generated features, contrasting to its other half, the DenseNet-Tiny-A. With that said, the DenseNet-Tiny-B produced a distinct set of shared features using the ImageNet and the CXR dataset, while the other DenseNet-Tiny-A preserved most of the ImageNet features on its upper layers and only had its lower dense layers trained with the CXR dataset.
Fig. 6

The transfer learning and partial layer freezing framework to train the Fused-DenseNet-Tiny.

The transfer learning and partial layer freezing framework to train the Fused-DenseNet-Tiny. With the two similar models having a different set of features fused, the proposed model's feature pool became robust even with fewer parameters involved.

Model compilation and training

Before the actual training process began, the model had its hyper-parameters and a loss function selected. Hyper-parameters represent the tunable parts of a DL model that can affect most of its learning procedure, which cannot be tweaked during training [28]. Together, an added loss function calculates and reduces errors during both the train and validation phases. For an efficient result, an appropriate selection of hyper-parameters and loss function is imperative. It is worth mentioning that, unlike other studies, this work did not perform any stringent optimization techniques to tune its hyper-parameters. Therefore, providing less assistance to train the model to prove its ease of reproducibility and adaptive ability towards an imported dataset.

Hyper-parameters

The tuned hyper-parameters of the model consisted of the Learning Rate (LR), Batch Size (BS), optimizer, Dropout Rate (DR), and epochs. In selection, the following values provided in Table 3 produced the best results during the experiments. A BS of 16 gave the model a fast yet non-exhausting training process combined with an Adam optimizer [29]. The selected optimizer recently had tremendous success and started to become a de facto algorithm to optimize most DCNN models that involved medical images [30]. Also, Adam attains a faster convergence with less memory consumption compared to the likes of a Stochastic Gradient Descent (SGD) [31] and RMSProp [32], making it a fitting candidate for this work. The LR had a value of 0.0001 that worked well with the rest of the selected hyper-parameters. Included, the DR of 0.5 provided an adequate regularization that reduced the size of dense units at random during learning that prevented the model from overfitting due to the robust flow of features [33].
Table 3

Selected hyper-parameters to train the Fused-DenseNet-Tiny.

Hyper-ParameterValue
LR0.0001
BS16
OptimizerAdam
DR0.5
Epochs25
Selected hyper-parameters to train the Fused-DenseNet-Tiny.

Loss function

Selecting an appropriate loss function helps the model identify errors efficiently by calculating the dissimilarities between a predicted and a ground truth class [34]. The proposed model with three CXR classes went with a Categorical Cross-Entropy (CCE) loss function that worked effectively with the Softmax classifier [35], considering it already falls under a multi-class task. In Eq. (1), M represents the three classes, Normal, COVID-19, and Pneumonia. Each class had its loss calculated separately upon each observation o from the ground truth label y that identified if its prediction p of sample c is correct [36].

Experimental results and discussion

In DL, proper evaluation using the appropriate metrics can determine how well a model performs according to a specific task. In this section, the Fused-DenseNet-Tiny undergoes a series of evaluations to validate its overall performance in diagnosing the three CXRs: Normal, COVID-19, and Pneumonia. As mentioned, this work used 20 % of the dataset or 1841 images for validation.

Confusion matrix

For a visual interpretation of how well the Fused-DenseNet-Tiny classified the validation samples individually, this work used a Confusion Matrix [37]. As illustrated in Fig. 7 , both confusion matrices consist of a True label or the ground truth, representing each image sample's labels, while the Predicted label represents the diagnosed samples. In which the CM uses a True Positive (T), True Negative (T), False Positive (F), and a False Negative (F) to identify whether the model had a correct or incorrect diagnosis of specific CXR. A T indicates that the model correctly diagnosed a COVID-19 or Pneumonia CXR. In contrast, a T implies that the model diagnosed a Normal CXR as Normal. On the other hand, the F means that the model diagnosed a Normal CXR as any of the two infections, while F diagnosed an infected CXR as Normal.
Fig. 7

The classification results of the Fused-DenseNet-Tiny visualized with a confusion matrix.

The classification results of the Fused-DenseNet-Tiny visualized with a confusion matrix. Upon evaluation, the Fused-DenseNet-Tiny had the most difficulty diagnosing Normal CXRs with 16 F CXRs, having 2.45 % of its entire set of samples diagnosed as Pneumonia. In terms of COVID-19, the model had a remarkable performance with only two F and one F, achieving 98.83 % correct diagnoses from all the COVID-19 CXRs. Simultaneously, the proposed model had 18 F of Pneumonia CXRs being Normal and having the model diagnose the entire 98.07 % of the Pneumonia CXRs correctly.

Sensitivity and specificity

From a medical standpoint, a sensitivity test serves as an essential metric to indicate the correctness of identifying a diseased person, while a specificity test pertains the other way around of identifying a person without a disease [38]. In DL, such a concept delivers a similar objective [39]. Illustrated in Fig. 8 , using a Receiver Operating Characteristic (ROC) curve and its Area Under the Curve (AUC), this work identified the performance of the Fused-DenseNet-Tiny based on sensitivity and specificity test in a varying threshold. ROCs with a larger AUC indicate better performance, whereas an AUC < 0.5 implies that the model merely conducted guesses and has no discrimination capability to diagnose the CXRs, making it unreliable [40].
Fig. 8

The Receiver Operating Characteristic and its Area Under the Curve.

The Receiver Operating Characteristic and its Area Under the Curve. As evaluated, the proposed model achieved a remarkable performance in sensitivity and specificity with a uniformed 1.00 AUC for all three classes of CXRs.

Precision and recall

In situations like this work with unbalanced dataset samples, a Precision-Recall (P-R) curve becomes a critical evaluation tool to determine F and F rates yielded through a discriminating threshold. In similar terms with the ROC, the P-R curve also uses an AUC for evaluation, whereas a smaller AUC indicates that the model is more prone to false predictions [41]. From the given illustration in Fig. 9 , the diagnosis for the Normal CXR had a 0.994 AUC, while the COVID-19 had 0.999, Pneumonia with 0.996, rendering a micro-average of 0.996 AUC. Even with the slight fluctuations from the unbalanced distribution of validation samples, the model still efficiently handled the diagnostic process as seen on the P–R curve.
Fig. 9

The Precision-Recall curve and its Area Under the Curve.

The Precision-Recall curve and its Area Under the Curve.

Saliency maps

For added transparency and visualization, this work used the Gradient-Weighted Class Activation Maps (GRAD-CAM) algorithm that functions by capturing a specific class's vital features from the last Conv layer of a CNN model to localize its important regions [42]. From the given samples in Fig. 10 , using the GRAD-CAM algorithm with the Fused-DenseNet-Tiny had its salient areas identified. The highlighted patched areas indicated the most salient features that the model believes to be important for diagnosing each CXR accordingly. The proposed model also shows that it did not rely on random guesses during its diagnosis. Though not precisely what some medical experts might expect, the GRAD-CAM still has its limitations, as it cannot entirely capture the desired areas without flaws [43]. Nonetheless, having an algorithm to visualize what CNNs like the Fused-DenseNet-Tiny sees emanates better trust and reliability. It is worth mentioning that the GRAD-CAM did not affect any of the proposed models' workability as it only served as a visualization tool for this work.
Fig. 10

Gradient-Weighted Class Activation Maps of the Fused-DenseNet-Tiny.

Gradient-Weighted Class Activation Maps of the Fused-DenseNet-Tiny.

Classification performance

For an overall identification of the Fused-DenseNet-Tiny's performance in diagnosing the three distinct CXRs, the use of the following metrics evaluated its critical points. The following includes the accuracy, precision, recall, and f1-score, calculated based on the following equations below [44]. Table 4 presents the overall performance of the Fused-DenseNet-Tiny in diagnosing 1841 samples of Normal, COVID-19, and Pneumonia CXRs. Upon evaluation, the Fused-DenseNet-Tiny achieved a reliable performance of 99.84 % accuracy with COVID-19, 98.10 % with Pneumonia, and 98.04 % with Normal CXRs. In terms of precision, recall, and f1-score, the Fused-DenseNet-Tiny model performed considerably well even with the unbalanced data and had no perceptible signs of bias or class superiority issues as reflected from the given and previous results.
Table 4

Performance of the trained Fused-DenseNet-Tiny in diagnosing the chest x-rays.

ClassesAccuracy (%)PrecisionRecallF1-scoreSample size
Normal98.040.980.970.97654
COVID-1999.840.991.000.99256
Pneumonia98.100.980.980.98931
Performance of the trained Fused-DenseNet-Tiny in diagnosing the chest x-rays.

Discussion

The Fused-DenseNet-Tiny had proven to perform efficiently, even with fewer parameters and less depth based on the evaluated results. However, to further showcase its contribution and improvements, this work compared its overall accuracy with other state-of-the-art models and studies that performed a similar diagnosis of COVID-19 in CXRs. From the presented results in Table 5 , the proposed Fused-DenseNet-Tiny achieved the highest accuracy of 97.99 % without the employment of hyper-parameter optimization, data augmentation, lengthy training epochs, ensembling, and any specialized hardware. It is worth mentioning that the proposed Fused-DenseNet-Tiny only performed under a standard GTX 1070 GPU (released 2016), an i5 4th generation CPU, and 16GB of RAM.
Table 5

Performance comparison of the proposed Fused-DenseNet-Tiny with other studies.

ModelAccuracy (%)ClassesType
Fused-DenseNet-Tiny (this work)97.99Normal, COVID-19, PneumoniaCXR
COVID-Net [7]93.30Normal, COVID-19, PneumoniaCXR
Modified ResNet-18 [8]96.37Normal, COVID-19, PneumoniaCXR
ECOVNet-EfficientNetB3 base [10]97.00Normal, COVID-19, PneumoniaCXR
Modified Xception [11]95.70Normal, COVID-19, PneumoniaCXR
DarkCovidNet [12]87.02Normal, COVID-19, PneumoniaCXR
DeTraC-ResNet18 [13]95.12Normal, COVID-19, SARSCXR
Hierarchical EfficientNetB3 [14]93.51Normal, COVID-19, PneumoniaCXR
Performance comparison of the proposed Fused-DenseNet-Tiny with other studies. Furthermore, this work also trained several state-of-the-art DCNNs with the same dataset. Table 6 shows that the proposed Fused-DenseNet-Tiny did not attain the highest overall performance rating due to its lesser and more straightforward feature learning process. However, compared with the significantly larger DenseNet121 that attained the highest accuracy of 98.48 %, the Fused-DenseNet-Tiny only falls 0.49 % less accurate with 97.99 %.
Table 6

Comparison of performance with other state-of-the-art models.

ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
DenseNet121 [18]98.4898.7198.5998.48
EfficientNetB0 [45]98.2198.5998.1898.39
Fused-DenseNet-Tiny (this work)97.9998.3898.1598.26
InceptionV3 [46]97.9998.3198.2398.26
ResNet152V2 [47]97.8898.2598.0998.17
Xception [48]97.6197.9297.8397.87
MobileNetV2 [49]97.1297.4697.7597.58
VGG16 [50]96.5897.0696.9496.97
InceptionResNetV2 [51]96.1494.4896.9095.59
Comparison of performance with other state-of-the-art models. Even though the Fused-DenseNet-Tiny fell short in attaining the highest accuracy in this work, in terms of the parameter size, the Fused-DenseNet-Tiny had shown an advantage over the rest of the presented state-of-the-art models having only around 1.2 million, as shown in Fig. 11 . Convincingly, the Fused-DenseNet-Tiny, even with a slightly minimal performance difference compared to InceptionV3, EfficientNetB0, and DenseNet121, still garnered a remarkable performance to size ratio.
Fig. 11

The parameter sizes of the Fused-DenseNet-Tiny and other state-of-the-art models.

The parameter sizes of the Fused-DenseNet-Tiny and other state-of-the-art models. As presented by the calculated results throughout this article, the proposed method to re-structure and train a state-of-the-art DCNN model like the DenseNet can considerably maintain most of its performance and save a large disk capacity and computing cost simultaneously. Such a solution can become easily deployable in smaller or lower-performing machines, including mobile devices, giving developing countries and locations that cannot confidently perform tests or diagnose CXRs infected by COVID-19 an opportunity to emanate better performance. Medical experts and the likes can gain more confidence in conducting their diagnosis and not require costly equipment through the proposed DL-based solution's assistance [52]. It is essential to know that even with the increase of training data, the model's weight size would not inflate compared to re-structuring and adding layers, making it conveniently scalable in most situations. Adding layers or depth also does not mean an increase in performance [53,54]. The Fused-DenseNet-Tiny and its small parameter size and dense structure can train faster even with more data than larger state-of-the-art models that train with lesser data, thus yielding immediate improvements from additional reliable data in the future.

Conclusion

Due to the relentless spread of COVID-19 infections, mass testing became an essential aspect of most people's lives today. However, the gold standard testing procedure like rRT-PCR requires specialized testing equipment and a trained medical practitioner. Even with a lesser substitute like CXRs, a rapid and less expensive method to rule out the infections of COVID-19 still induces complexity in most underdeveloped countries. Hence, people started to automate such a difficult task through DL. This work served as an additional contribution with its lightweight yet efficient design that requires less effort to reproduce and does not require expensive equipment to help diagnose COVID-19 infections from CXRs. As a result, the proposed model yielded a slight performance improvement but a massive decrease in computing cost and parameter size over other state-of-the-art models and existing studies. However, even with its lightweight design, a specific caveat still shows that the Fused-DenseNet-Tiny cannot outperform its larger counterpart due to its reduced extraction capabilities. Though not far in terms of performance, this work hypothesizes that hyper-parameter optimization, and the potential addition of more data may yield additional improvements that can alleviate such a problem in the future. Nonetheless, even with minimal shortcomings, this work still concludes that the fusion of a mirrored truncated DenseNet, with its equivalent partially trained with ImageNet features and the other with shared weights from the CXR and ImageNet, the proposed model still performed efficiently even with less computing cost, data, and dependence to other sophisticated optimization methods towards the diagnosis of COVID-19 from CXRs. Furthermore, the model can still induce further improvements upon applying the mentioned methods above, besides adding more data.

Authorship statement

I, Francis Jesmar P. Montalbo as the sole author of this paper, dedicated myself and time to the following. Conceptualizing the study. Acquiring, handling, and analyzing data. Performing experiments, testing, and evaluating results. Illustrating figures. Drafting, revising, and finalizing the manuscript. With this document, I certify and guarantee that this material has not been or will not be submitted or published in any other journal. This document was solely submitted to the Biomedical Signal Processing and Control journal with an ISSN of 1746–8094.

Ethical procedure

This research article meets all applicable standards with regards to the ethics of experimentation and research integrity, and the following is being certified/declared true. As a researcher in my field of specialization, the paper has been submitted with full responsibility, following the due ethical procedure, and there is no duplicate publication, fraud, plagiarism, or concerns about animal or human experimentation.

Ethical approval

This work did not involve humans, animals, and other living specimens during experiments.

Code and data availability

The author provides the links to the source code and dataset used in this work → https://github.com/francismontalbo/fused-densenet-tiny

Dataset and code availability

For ease of evaluation, this work provides both the dataset source and code required to simulate the proposed model. https://github.com/francismontalbo/fused-densenet-tiny.
  13 in total

Review 1.  Receiver Operating Characteristic (ROC) Curve Analysis for Medical Diagnostic Test Evaluation.

Authors:  Karimollah Hajian-Tilaki
Journal:  Caspian J Intern Med       Date:  2013

2.  Facing Imbalanced Data Recommendations for the Use of Performance Metrics.

Authors:  László A Jeni; Jeffrey F Cohn; Fernando De La Torre
Journal:  Int Conf Affect Comput Intell Interact Workshops       Date:  2013

3.  COVID-19 identification in chest X-ray images on flat and hierarchical classification scenarios.

Authors:  Rodolfo M Pereira; Diego Bertolini; Lucas O Teixeira; Carlos N Silla; Yandre M G Costa
Journal:  Comput Methods Programs Biomed       Date:  2020-05-08       Impact factor: 5.428

4.  Chest X-rays findings in COVID 19 patients at a University Teaching Hospital - A descriptive study.

Authors:  Misbah Durrani; Inam Ul Haq; Ume Kalsoom; Anum Yousaf
Journal:  Pak J Med Sci       Date:  2020-05       Impact factor: 1.088

5.  Automated detection of COVID-19 cases using deep neural networks with X-ray images.

Authors:  Tulin Ozturk; Muhammed Talo; Eylul Azra Yildirim; Ulas Baran Baloglu; Ozal Yildirim; U Rajendra Acharya
Journal:  Comput Biol Med       Date:  2020-04-28       Impact factor: 4.589

6.  CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest x-ray images.

Authors:  Asif Iqbal Khan; Junaid Latief Shah; Mohammad Mudasir Bhat
Journal:  Comput Methods Programs Biomed       Date:  2020-06-05       Impact factor: 5.428

7.  The continuing 2019-nCoV epidemic threat of novel coronaviruses to global health - The latest 2019 novel coronavirus outbreak in Wuhan, China.

Authors:  David S Hui; Esam I Azhar; Tariq A Madani; Francine Ntoumi; Richard Kock; Osman Dar; Giuseppe Ippolito; Timothy D Mchugh; Ziad A Memish; Christian Drosten; Alimuddin Zumla; Eskild Petersen
Journal:  Int J Infect Dis       Date:  2020-01-14       Impact factor: 3.623

Review 8.  Coronavirus Disease (COVID-19): Spectrum of CT Findings and Temporal Progression of the Disease.

Authors:  Mingzhi Li; Pinggui Lei; Bingliang Zeng; Zongliang Li; Peng Yu; Bing Fan; Chuanhong Wang; Zicong Li; Jian Zhou; Shaobo Hu; Hao Liu
Journal:  Acad Radiol       Date:  2020-03-20       Impact factor: 3.173

9.  Evaluation of Scalability and Degree of Fine-Tuning of Deep Convolutional Neural Networks for COVID-19 Screening on Chest X-ray Images Using Explainable Deep-Learning Algorithm.

Authors:  Ki-Sun Lee; Jae Young Kim; Eun-Tae Jeon; Won Suk Choi; Nan Hee Kim; Ki Yeol Lee
Journal:  J Pers Med       Date:  2020-11-07
View more
  7 in total

1.  COVLIAS 2.0-cXAI: Cloud-Based Explainable Deep Learning System for COVID-19 Lesion Localization in Computed Tomography Scans.

Authors:  Jasjit S Suri; Sushant Agarwal; Gian Luca Chabert; Alessandro Carriero; Alessio Paschè; Pietro S C Danna; Luca Saba; Armin Mehmedović; Gavino Faa; Inder M Singh; Monika Turk; Paramjit S Chadha; Amer M Johri; Narendra N Khanna; Sophie Mavrogeni; John R Laird; Gyan Pareek; Martin Miner; David W Sobel; Antonella Balestrieri; Petros P Sfikakis; George Tsoulfas; Athanasios D Protogerou; Durga Prasanna Misra; Vikas Agarwal; George D Kitas; Jagjit S Teji; Mustafa Al-Maini; Surinder K Dhanjil; Andrew Nicolaides; Aditya Sharma; Vijay Rathore; Mostafa Fatemi; Azra Alizad; Pudukode R Krishnan; Ferenc Nagy; Zoltan Ruzsa; Mostafa M Fouda; Subbaram Naidu; Klaudija Viskovic; Mannudeep K Kalra
Journal:  Diagnostics (Basel)       Date:  2022-06-16

2.  Truncating a densely connected convolutional neural network with partial layer freezing and feature fusion for diagnosing COVID-19 from chest X-rays.

Authors:  Francis Jesmar P Montalbo
Journal:  MethodsX       Date:  2021-06-05

3.  Truncating fined-tuned vision-based models to lightweight deployable diagnostic tools for SARS-CoV-2 infected chest X-rays and CT-scans.

Authors:  Francis Jesmar Montalbo
Journal:  Multimed Tools Appl       Date:  2022-03-03       Impact factor: 2.577

4.  Study on the Grading Model of Hepatic Steatosis Based on Improved DenseNet.

Authors:  Ruwen Yang; Yaru Zhou; Weiwei Liu; Hongtao Shang
Journal:  J Healthc Eng       Date:  2022-03-17       Impact factor: 2.682

5.  Deep feature fusion classification network (DFFCNet): Towards accurate diagnosis of COVID-19 using chest X-rays images.

Authors:  Jingyao Liu; Wanchun Sun; Xuehua Zhao; Jiashi Zhao; Zhengang Jiang
Journal:  Biomed Signal Process Control       Date:  2022-04-13       Impact factor: 5.076

6.  DS-MENet for the classification of citrus disease.

Authors:  Xuyao Liu; Yaowen Hu; Guoxiong Zhou; Weiwei Cai; Mingfang He; Jialei Zhan; Yahui Hu; Liujun Li
Journal:  Front Plant Sci       Date:  2022-07-22       Impact factor: 6.627

7.  Automatic coronavirus disease 2019 diagnosis based on chest radiography and deep learning - Success story or dataset bias?

Authors:  Jennifer Dhont; Cecile Wolfs; Frank Verhaegen
Journal:  Med Phys       Date:  2022-01-12       Impact factor: 4.506

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.