Literature DB >> 34141886

An optimized transfer learning-based approach for automatic diagnosis of COVID-19 from chest x-ray images.

Waleed M Bahgat1, Hossam Magdy Balaha2, Yousry AbdulAzeem3, Mahmoud M Badawy2.   

Abstract

Accurate and fast detection of COVID-19 patients is crucial to control this pandemic. Due to the scarcity of COVID-19 testing kits, especially in developing countries, there is a crucial need to rely on alternative diagnosis methods. Deep learning architectures built on image modalities can speed up the COVID-19 pneumonia classification from other types of pneumonia. The transfer learning approach is better suited to automatically detect COVID-19 cases due to the limited availability of medical images. This paper introduces an Optimized Transfer Learning-based Approach for Automatic Detection of COVID-19 (OTLD-COVID-19) that applies an optimization algorithm to twelve CNN architectures to diagnose COVID-19 cases using chest x-ray images. The OTLD-COVID-19 approach adapts Manta-Ray Foraging Optimization (MRFO) algorithm to optimize the network hyperparameters' values of the CNN architectures to improve their classification performance. The proposed dataset is collected from eight different public datasets to classify 4-class cases (COVID-19, pneumonia bacterial, pneumonia viral, and normal). The experimental result showed that DenseNet121 optimized architecture achieves the best performance. The evaluation results based on Loss, Accuracy, F1-score, Precision, Recall, Specificity, AUC, Sensitivity, IoU, and Dice values reached 0.0523, 98.47%, 0.9849, 98.50%, 98.47%, 99.50%, 0.9983, 0.9847, 0.9860, and 0.9879 respectively.
© 2021 M. Bahgat et al.

Entities:  

Keywords:  COVID-19; Classification; Deep convolutional neural network; Transfer learning; X-ray images

Year:  2021        PMID: 34141886      PMCID: PMC8176553          DOI: 10.7717/peerj-cs.555

Source DB:  PubMed          Journal:  PeerJ Comput Sci        ISSN: 2376-5992


Introduction

Coronavirus Disease 2019 (COVID-19) had been detected in Wuhan, China, at the end of the year 2019 and represented a severe health issue worldwide. The recent Coronavirus (COVID-19) has been declared a pandemic by the World Health Organization (WHO) in March 2020. Mankind faces many pandemics like Spanish flu in 1918, Severe Acute Respiratory Syndrome (SARS) in 2003, and presently COVID-19. These infections are airborne and might, therefore, promptly transmittable taint expansive bunches of people (Ghayvat et al., 2021). There is no crisis within the history of mankind that comes anyplace close to the scale of the COVID-19 widespread, with such numerous nations being influenced at the identical time in a very brief time (World Health Organization, 2020a). The infodemic that the widespread of such a scale has caused makes it troublesome for people to get solid direction around COVID-19 avoidance and remedy (Choudrie et al., 2021). Unfortunately, the spread of COVID-19 is exponentially, and the transmission process is not clearly understood. Today, dealing with Coronavirus is an important healthcare challenge for humanity all over the world. Even though the Coronavirus infection in common incorporates small or no side effects, it causes lethal pneumonia in 2:8% of the infected patients (Hoehl et al., 2020; Lu et al., 2020; Sohrabi et al., 2020). Normally it takes 5:6 days from the disease day with the infection for indications to appear. However, it may take up to 14 days in some cases. As shown in Fig. 1, without rapid testing, the number of COVID-19 cases could increase beyond the healthcare system’s capacity to deal with serious cases (red curve in relation to healthcare system capacity line). Fast and accurate detection reduces the daily rate of new cases (flattening the blue curve).
Figure 1

Coronavirus outbreaks and lowering infection rates (World Health Organization, 2020b).

It is vital to distinguish (i.e., diagnose) the COVID-19 as quickly as conceivable and in an exact way to diminish the fast spread of the infection. In general, three different methodologies, as seen in Fig. 2, can be used to recognize COVID-19 patients as follows: (1) RealTime reverse transcriptase-polymerase Chain Reaction (RT-PCR), (2) Chest Computed Tomography (CT) imaging scan, and (3) Numerical laboratory tests. The current common strategy to recognize COVID-19 patients is the RT-PCR test (Ai et al., 2020) which is fairly sensitive and reliable. Unfortunately, depending only on the RT-PCR causes a remarkable delay in the diagnosis of the suspected patients. Besides, utilizing the RT-PCR test has numerous troubles within the extremely hit regions, particularly amid early flare-ups of this disease. On the other hand, the high false-negative rates are a true challenge for the test labs. Many factors affect the sample result, such as test planning and quality control (Liang, 2020).
Figure 2

Coronavirus diagnose techniques.

As a result, chest imaging, such as chest CT or chest X-ray (CXR), is utilized as a first-line examination to detect the virus infection (Rubin et al., 2020; Wong et al., 2020). Chest imaging technologies, especially CXR, are broadly accessible and economical. For this reason, radiologists use chest imaging to detect early lesions in the lung at high speed and sensitivity. Concerning the COVID-19, several aspects of lung anomaly such as bilateral abnormalities, interstitial abnormalities, lung consolidation, and ground-glass opacities are showed in chest images (Guan et al., 2020). Consequently, examining the suspected patients’ chest images presents an essential part remarkable potential for screening processes, and early determination of COVID-19 disease (Kandaswamy et al., 2014). Unfortunately, the diagnosis process mainly relies on the radiologists’ visual diagnosis, which leads to many issues. At first, it takes an exceptionally long time to diagnose since chest imaging contains hundreds of slices, which takes a long time to analyze. Much other pneumonia has similar aspects to COVID-19, so, only radiologists with accumulative experience to diagnose COVID-19 disease. The artificial intelligence (AI) branch, especially deep learning (DL), has been used to automatically identify lung diseases in medical imaging with significantly diagnostic accuracy (Guan et al., 2020; Rubin et al., 2020). Deep learning is efficient in dealing with the medical dataset, particularly those datasets containing a huge number of training samples. Recently, many research papers address the detection of COVID-19 pneumonia and classify the severity of COVID-19. These research studies try to automatically detect the infected patients early to help society by isolating them to prevent or decrease the native spread. Deep transfer learning (DTL) is a deep learning approach trained on a source problem and then reused to solve a target problem. DTL can be used for two reasons: (1) the transfer of the extracted features from the source issue (Kandaswamy et al., 2014) and (2) the retrain of only the network unfrozen layers by constraining not to over train for the target mission (Yosinski et al., 2014). Employing Convolution Neural networks (CNN) using Transfer Learning is recently used to achieve very promising approaches to diagnose several diseases such as cancer (Bhatt, Ganatra & Kotecha, 2021) and Parkinson (Suzuki, 2017). Although the research applied in convolutional neural network architectures achieved promising results to detect the COVID-19, the huge number of existing hyperparameters still represents a challenge to achieve better and promising results (Ng et al., 2020). There are recent investigations to optimize CNN’s hyperparameters by utilizing metaheuristics. The hyperparameters of the CNN network, such as learning rate, kernel size, and kernel type of convolutional blocks, should be tuned to enhance the network’s performance. Zhao, Zhang & Wang (2020) proposed the Manta Ray Foraging Optimization (MRFO) algorithm, inspired by manta rays’ intelligent behaviors. MRFO overcomes the major drawbacks of well-known metaheuristics algorithms, such as slow search speed and easy premature convergence. MRFO outperforms the other optimization algorithms to search precision, convergence rate, stability, and local optimal value avoidance (Wong et al., 2020). This study suggests using the MRFO algorithm to fine-tune the CNN network’s hyperparameters to achieve more accurate results. This study’s main objective is to introduce an Optimized Transfer Learning-based Approach for Automatic Detection of COVID-19 (OTLD-COVID-19). The OTLD-COVID-19 approach performs the classification and recognition of the lung X-ray images for COVID-19 diseases. The OTLD-COVID-19 approach uses CNN and MRFO algorithm for the parameters and hyperparameters optimizations, respectively. The rest of the paper is structured as follows: “Background” represents the CNNs and MRFO Algorithm theoretical background. “Related Work” discusses the related work. “The Proposed OTLD-COVID-19 Approach” shows the proposed OTLD-COVID-19 approach details. The results, achieved by the proposed technique, are presented in “Experiments, Results and Discussion”. Finally, the conclusions and future work are drawn.

Paper contributions

In this paper, an approach for automatically detecting COVID-19 using transfer learning is proposed to achieve diagnosis reliably. The OTLD-COVID19 approach works on lung X-ray images to achieve both recognition and classification of COVID-19 diseases. This approach aims to achieve high performance in both processes. To that end, it seeks the best hyperparameters combination to optimize CNN’s parameters. There are many meta-heuristics optimization algorithms with different approaches, including try-and-error, deterministic, and stochastic. This study uses a meta-heuristic optimizer to search the scope automatically. In this vein, to improve the classification performance, the OTLD-COVID-19 approach utilized CNN and MRFO algorithm for the automatic parameters and hyperparameters optimization respectively. The OTLD-COVID-19 approach is used as a pillar to improve overall accuracy and reduce computational cost. The contributions of the current study are summarized in points as follow: Proposing the OTLD-COVID-19 approach for rapid COVID-19 diagnosis using automated X-ray image processing. Twelve popular CNN architectures, with four classes (“COVID-19,” “Bacterial Pneumonia,” “Viral Pneumonia,” and “Normal”), are used in the transfer learning technique for the detection of COVID-19 patients. The Manta-Ray Foraging Optimization (MRFO) algorithm is used to optimize the network hyperparameters’ values to enhance the classification performance. The proposed technique is adaptable and scalable; there is no need to manually assign the CNN architecture’s hyperparameters values. Regularization, optimization, dropout, and data augmentation are studied through the different reported experiments. The proposed technique is compared with the other state-of-the-art techniques and studies. The achieved results of the standard performance metrics are very promising. For the convenience of readers, Fig. 3 depicts the guidelines of the four primary sections that reflect the contributions of this study. First, “The Proposed OTLD-COVID-19 Approach” introduces the proposed OTLD-COVID-19 approach. The OTLD-COVID-19 approach is characterized by its adaptability and scalability. It consists of five phases: (1) Acquisition phase, (2) Preprocessing phase, (3) Augmentation phase, (4) Training, Classification, and Optimization (TCO) phase, and (5) Deployment phase. Second, “The Proposed OTLD-COVID-19 Approach” illustrates the MRFO algorithm, which uses various foraging mathematical models to optimize the key hyperparameters’ values automatically. Third, the deployment phase of the proposed approach uses computed hyperparameters to build a diagnostic model. Each model evaluates the COVID-19 dataset to classify the cases into the main four categories (i.e., “COVID-19,” “Bacterial Pneumonia,” “Viral Pneumonia,” and “Normal”). Fourth, in “Experiments, Results and Discussion”, different reported experiments are introduced. The achieved results of the standard performance metrics are very promising compared with the other state-of-the-art techniques and studies.
Figure 3

The guidelines that represent the contributions of this study.

In a summary, our OTLD approach improves the network performance in several ways. First, due to the lack of COVID 19 data, the proposed dataset is created from eight different datasets to increase the amount of data and avoid overfitting. Second, the MRFO optimization algorithm is used to obtain the optimal values of network hyperparameters. MRFO overcomes the main drawbacks of well-known metaheuristic algorithms, such as slow search speed and slight premature convergence. MRFO outperforms the other optimization algorithms in terms of search precision, convergence rate, stability, and local optimal value avoidance. Third, Transfer Learning (TL) is used to achieve the best network performance. Here, the MRFO optimization algorithm is applied to twelve different CNN architectures to obtain the best combination of hyperparameters for each architecture and obtain the architecture with the best performance. Finally, many performance metrics are measured to ensure the competence of the architecture with the best performance.

Background

CNN architectures

There is a continuous research investigation in the CNN architectures, and, notably, the noteworthy achievement in the CNN performance happened from 2015 to 2019. Khan et al. (2019) classified the CNN architectures into seven main categories. In this section, the most applicable CNN architectures from the mentioned classes are described. AlexNet (Antonellis et al., 2015) is designed to be the first known deep CNN architecture. AlexNet increases CNN’s depth by applying many different parameter optimization strategies to improve the learning capacity of the CNN (Antonellis et al., 2015). Although increasing the depth enhanced the generalization for different image resolutions, it caused the network to suffer from overfitting problems. Krizhevsky, Sutskever & Hinton (2012) adopted the idea of Hinton (Dahl, Sainath & Hinton, 2013; Srivastava et al., 2014) to solve the overfitting problem. They enforced the model to learn more robust features through skipping some transformational units randomly during the training process. Moreover, they used the Rectified Linear Unit (ReLU) activation function to enhance the convergence rate by mitigating the vanishing gradient issue (Nair & Hinton, 2010). Simonyan & Zisserman (2014) proposed VGG, a CNN architecture designed simply and efficiently. It has 19 layers deeper than AlexNet to simulate the relationship between the network depth and capacity (Antonellis et al., 2015; Hodan, 1954). VGG addressed the large-size filter effect by replacing the large-size filter with a stack of (3 × 3) filters. Using the small-size filters enhanced the computation complexity. Unfortunately, the VGG still had a huge number of parameters, leading to severe difficulties in deploying it on low resource systems. Wu, Zhong & Liu (2018) proposed the CNN’s residual learning principle and efficiently trained the deep networks. ResNet presented a deeper network with less computational complexity than the previously proposed networks. ResNet is deeper than AlexNet and VGG by 20 and eight times, respectively (Simonyan & Zisserman, 2014). The GoogleNet architecture was designed mainly to enhance the accuracy by reducing the computational cost (Szegedy et al., 2015). For this reason, the CNN inception block principle is presented where the conventional layers are replaced in small blocks (Lin, Chen & Yan, 2013). Each block has different size filters to get the spatial data at diverse scales. Inception-V3, Inception-V4, and Inception-ResNet are modified and enhanced versions of Inception-V1 and Inception-V2 (Roos & Schuttelaars, 2015; Szegedy et al., 2015; Szegedy et al., 2016). The main concept of Inception-V3 is to reduce the computational cost of deep networks without influencing the generalization. For that reason, Szegedy et al. (2016) used small and asymmetric filters ((1 × 5) and (1 × 7)) instead of large size filters ((5 × 5) and (7 × 7)). Besides, they used (1 × 1) convolution as a bottleneck in the front of the large filters. In Inception-ResNet, Szegedy et al. (2016) combined both the power of residual learning and inception block. DenseNet was mainly designed to handle the vanishing gradient issue. ResNet specifically holds information through additive identity transformations; it suffers from a series problem, resulting in several layers that can contribute very little or no information (Blei, Ng & Jordan, 2003). DenseNet employed cross-layer connectivity differently. A feed-forward fashion is used to connect each previous layer to the next coming layer. The network’s information flow is substantially improved as DenseNet used the loss function to grant each layer direct access to the gradients. Xception is known to be an extreme Inception architecture, where the concept of a depth-wise separable convolution is embraced (Chollet, 2017) Xception achieved significant enhancements since it broadened the original inception block and regulated the computational complexity. The different spatial dimensions ((1 × 1), (5 × 5), and (3 × 3)) are replaced by a single dimension (3 × 3) followed by a (1 × 1) convolution. Also, Xception convolved each feature map in a separate way to get easier computations. ResNet extended and improved the Inception Network (Xie et al., 2017). The authors applied the concept of splitting, transforming, and merging efficiently and simply. Besides, the cardinality concept is introduced (Szegedy et al., 2015). Cardinality is an extra dimension, which refers to the size of the set of transformations (Han et al., 2018; Sharma & Muttoo, 2018). The Inception network enhanced the convolution CNNs learning capability and ensured the efficient use of network resources. MobileNet (Howard et al., 2017) is a recent class of deep learning architectures explicitly designed for quick inference on mobile devices. MobileNets and other conventional models’ key difference is that two more efficient stages are added than the standard convolutional operation is decomposed into two more efficient stages. The depthwise separable convolutions are used to perform a single convolution on each color channel rather than combining all three and flattening it. In MobileNet V1, the pointwise convolution either kept the number of channels the same or doubled them. At the same time, MobileNetV2 (Sandler et al., 2018) decreases the number of channels. Table 1 compares the various CNN architectures.
Table 1

The CNN architectures comparison.

ArchitectureParametersError rateCategory
AlexNet60 MImageNet: 16.4Spatial Exploitation
VGG138 MImageNet: 7.3Spatial Exploitation
GoogLeNet4 MImageNet: 6.7Spatial Exploitation
ImageNet: 3.5
Inception-V323.6 MMulti-Crop: 3.58Depth + Width
Single-Crop: 5.6
Inception-V435 MImageNet: 4.01Depth +Width
Inception-ResNet55.8 MImageNet: 3.52Depth + Width + Multi-Path
ResNet25.6 MImageNet: 3.6Depth + Multi-Path
1.7 MCIFAR-10: 6.43
Xception22.8 MImageNet: 0.055Width
CIFAR-10: 3.58
CIFAR-100: 17.31
ResNeXt68.1 MImageNet: 4.4Width
25.6 MCIFAR-10+: 3.46
DenseNet25.6 MCIFAR100+: 17.18Multi-Path
15.3 MCIFAR-10: 5.19
15.3 MCIFAR-100: 19.64
MobileNet-V14.2 MImageNet: 10.5Depth + Width
MobileNet-V23.5 MImageNetDepth + Width

Transfer learning and CNN hyperparameters

Transfer learning is an effective representation learning approach in which the learned knowledge gained from a certain mission is used to enhance generalization about another. Transfer learning is much recommended when the number of images from the available datasets is relatively small. In this case, the original architecture and its weights are preserved and can be reused, especially when the used dataset in training the original architecture is vast. For example, the network architectures trained on the ImageNet dataset such as VGG (Simonyan & Zisserman, 2014) and DenseNet (Blei, Ng & Jordan, 2003) are extremely useful in medical image processing since it keeps the features of medical images in the ImageNet dataset. There are two common strategies to apply transfer learning: feature extraction and fine-tuning. In the feature extraction strategy, the last feed-forward layer(s) of the network is frozen. So, not all the weights are optimized; only the newly added layers are optimized during training. In the fine-tuning strategy, the pre-trained network is used as a starting point, and none of the weights are frozen, so all the network weights are optimized for the new task (Alshazly et al., 2019; Azizpour et al., 2015). When a fine-tuning strategy is adopted, it is recommended to apply lower learning rates to the pre-trained network to avoid the initial weights’ destruction. Hyperparameters tuning is essential since it controls the overall behavior of a learning model. The main objective of tuning the hyperparameters is to get an optimal combination of hyperparameters that minimizes a predefined loss function to improve the network performance. CNN hyperparameters are classified into two main categories. The hyperparameters related to the network structure and The hyperparameters that determine how the network is trained. The main hyperparameters of both categories are described in Table 2.
Table 2

The CNN hyperparameters.

CategoryHyperparametersDefinition
Network StructureHidden layersIt represents the number of layers between the input and output layer.
Kernel SizeIt indicates the height and width of the 2D convolution window.
Kernel TypeIt specifies the applied filter (e.g. edge detection, sharpen).
StrideIt specifies the step size of the kernel when crossing the image.
PaddingThe extra pixels of filler around the boundary of the input image that are set to zero.
DropoutIt defines the percentage of neurons that should be ignored to prevent overfitting.
Activation FunctionsThey are the mathematical equations that allow the model to learn nonlinear prediction boundaries.
Training MethodologyLearning RateIt defines how quickly a network updates its parameters.
MomentumIt specifies the value to let the previous update affect the current weight update.
The Epochs NumberThe number of iteration when the dataset is trained.
Batch SizeIt defines the number of patterns applied to the network before the weights are updated.
OptimizerIt defines the parameters updating technique.
Several methods are used to determine the hyperparameters’ values, including manual search, grid search, random search, and optimization techniques. This paper proposes using the MRFO optimization technique to find out the values of the hyperparameters.

Manta ray foraging optimization

Manta Ray Foraging Optimization (MRFO) (Zhao, Zhang & Wang, 2020) is a swarm meta-heuristic optimization algorithm bio-inspired by a foraging strategy practiced by the manta rays to capture the prey. Manta rays have developed various powerful and intelligent foraging strategies, such as chain foraging, cyclone foraging, and somersault foraging. Chain foraging mimics the intrinsic behavior of the food search. Foraging manta rays line up in an organized fashion to capture lost preys missed or unnoticed by the chain’s previous manta rays. This cooperative interaction between rival manta rays decreases the risk of prey loss while also increasing food rewards. Cyclone foraging occurs when there is a high density of prey. The tail end of the manta ray connects with its head forming a spiral to produce a vertex in the eye of a cyclone, causing the filtered water to rise to the surface. This complex mechanism allows a manta ray to capture the prey easily. The last foraging strategy is the somersault foraging. When manta rays find a food source (plankton), they perform backward somersaults before circling around the plankton, pulling it towards them. These foraging behaviors are extremely successful, despite their rarity in nature. The following sections cover the mathematical models for each foraging strategy. It worth mentioning that the used symbols in the current study for the MRFO models are similar to the MRFO original paper to avoid misunderstandings.

Chain foraging

Manta rays hunt for prey plankton and swim towards it after determining its location. The best location is one with a higher plankton concentration. Manta rays form a foraging chain by forming a line from head to tail. Each manta ray adjusts its position based on the best solution achieved so far and the location of the one in the front. Figure 4 shows the chain behavior.
Figure 4

Chain foraging behavior in the 2-D space illustration (Zhao, Zhang & Wang, 2020).

Cyclone foraging

When a group of manta rays recognizes a plankton group in deep water, they form a foraging chain and make spiral movements as they approach the food source. During the cyclone foraging process, flocked manta rays not only pursue the manta ray in front of them to ensure the chain’s continuity, but they also chase a spiral pathway to get to the target prey. Figure 5 shows the cyclone behavior.
Figure 5

Cyclone foraging behavior in the 2-D space illustration (Zhao, Zhang & Wang, 2020).

Somersault foraging

This foraging scheme considers the best prey location as a pivot point. Each manta-ray in the population searches around this point to migrate to a new location in the search domain. In the search space, all individuals progressively approximate to the optimal solution. As a result, the range of somersault foraging is reduced as iterations increase. Figure 6 demonstrates the pattern of somersault foraging in MRFO.
Figure 6

Somersault foraging behavior in the MRFO illustration (Zhao, Zhang & Wang, 2020).

Related work

Convolutional Neural Network (CNN) is considered one of the most effective deep learning approaches for accurately analyzing medical images. The main features to identify the COVID-19 in medical images include bilateral distribution of patchy shadows and ground-glass opacity (Wang et al., 2020). The research effort in the COVID-19 detection can be classified into three different perspectives related to deep learning techniques. These perspectives are tailored to CNN, transfer learning, and hybrid architectures. This section discusses the research studies to automate the detection of the COVID-19 according to these perspectives. A tailored CNN Architectures is a CNN network designed specifically to detect the COVID-19. The network is trained for the time. Mukherjee et al. (2020) introduced introduced a shallow CNN architecture consisting of four layers. Light architecture is to minimize the number of parameters (i.e., weights) to speed up computational time. Besides, the shallow (or lightweight) CNN design avoids possible overfitting that faces architectures with heavy usage of parameters. The shallow CNN architecture is well fit for mass population screening, especially in resource-constrained areas. The tailored shallow CNN model is designed to diagnose 2 class classification (“COVID” and “Normal”). It achieved an accuracy, sensitivity, and (Area Under Curve) AUC of 96.92%, 0.942, and 0.9869 respectively. Wang, Lin & Wong (2020) introduced COVID-Net, a tailored deep CNN architecture for detecting the COVID-19 cases using CXR images. They claimed that COVID-Net is one of the first open-source network designs for COVID-19 detection. They also introduced COVIDx, an open-access benchmark dataset with 13,975 CXR images representing 13,870 patient cases. The COVIDx dataset is generated by combining and modifying five open access existing datasets having chest scans. Their experimental results showed the achieved accuracy is 92.4% to classify “Normal,” “non-COVID Pneumonia,” and “COVID-19” classes. Majeed et al. (2020) introduced CNN-X, a tailored CNN architecture that holds four parallel layers. Each layer has 16 filters in three different sizes (3 × 3), (5 × 5), and (9 × 9). (3 × 3) filters detect the local-features while global features are detected by (9 × 9) filter. The (5 × 5) filter is used to detect what is missed by the other two filters. Then, the convolved image is applied to batch normalization and a ReLU activation function. Afterward, average pooling and maximum pooling are applied. The reason for using different size filters is to detect. They used Two COVID-19 X-ray image datasets in addition to a large dataset for other infections. Hammoudi et al. (2020) proposed a tailored CNN architecture to handle the two data modalities (CT and X-rays). Their model consists of nine layers for detecting COVID-19 positive cases. They trained and tested their network using both CT scans and X-rays. The experimental results show that their architecture achieved an overall accuracy of 96.28%. Hussain et al. (2021) presented a CoroDet, a tailored 22-layer CNN architecture to detect COVID-19 using both chest X-ray and CT modalities. The architecture consists of several layers: convolution layer, pooling layer, dense layer, flatten layer, and three activation functions. The CoroDet is designed to diagnose 2 class classification (“COVID-19” and “Normal”), 3 class classification (“COVID-19”, “Normal”, and “Pneumonia”), and 4 class classification (“COVID-19”, “Normal”, “non-COVID Viral Pneumonia”, and “non-COVID Bacterial Pneumonia”). Their architecture’s accuracy varies from 99.1% for the two classes to 91.2% for the four classes classification. Traditional transfer learning strategies are promising since the COVID-19 pneumonia CXR data is very limited. In this work, the popular deep learning architectures are customized to the purpose of the COVID-19 detection. Only the last few layers of the pre-trained model are replaced and retrained. The modified CNN architecture gets the advantages from the base CNN. The learned feature representations are fine-tuned to improve performance. Farooq & Hafeez (2020) suggested using the ResNet50 architected. They replaced the head of the trained model with another head containing a sequence of Adaptive average/max pooling, batch normalization, drop out, and linear. The ResNet50 weights are pre-trained using the ImageNet dataset that has X-ray images with different sizes. Apostolopoulos & Mpesiana (2020) tested five standard CNN architectures, including VGG19, InceptionNet, MobileNetV2, XceptionNet, and Inception. Different hyperparameters are used to identify the COVID-19. They used two different datasets having X-ray images from public medical repositories. Their results showed that the best-achieved accuracy, sensitivity, and specificity are 96.78%, 98.66%, and 96.46%, respectively, obtained from MobileNetV2 architecture. Narin, Kaya & Pamuk (2020) used five pre-trained convolutional neural network-based models to identify the COVID-19 using chest X-ray images. They implemented three different binary classifications with four classes (“COVID-19”, “Normal” (i.e., “Healthy”), “Viral Pneumonia,” and “Bacterial Pneumonia”). The results showed that the pre-trained ResNet50 model achieves the best classification performance. Islam et al. (2020) performed modality-specific transfer learning through retraining the ImageNet Network on the RSNA CXR collection to learn CXR modality-specific features and detect the abnormality. The used collection contains both normal CXRs and abnormal images having pneumonia-related opacities. Dropout is used to overcome overfitting where the regularization is restricted, and generalization is improved by reducing the model sensitivity to the training input’s specifics. The different hyperparameters of the CNNs are optimized using a randomized grid search method. Nayak et al. (2021) introduced a deep Learning architecture to automate the COVID-19 detection using X-ray images. They evaluated the performance of eight CBB architectures to detect COVID-19 cases. The introduced architectures are compared by considering several different hyperparameters values. The results showed that the ResNet-34 model achieved a higher accuracy of 98.33%. Khan, Shah & Bhat (2020) presented CoroNet architecture, a pre-trained CNN architecture to detect COVID-19 pneumonia from three different kinds of pneumonia using CXR images. The CNN architecture relies on Exception (Extreme Inception) and contains 71 layers trained on the ImageNet dataset. The authors introduced a balanced dataset containing 310 normal, 330 bacterial, 327 virals, and 284 COVID-19 resized CXR images. The experimental results showed that the CoroNet architecture achieved an accuracy of 0.87 and an F1-score of 0.93 for the COVID-19 detection. In most typical deep learning architectures, the CNN is used for both feature extraction and classification. Combined architectures use CNN either for feature extraction and apply another classifier to identify the COVID-19 patients or classify and use other algorithms to extract and optimize features. The hybrid architecture combines different deep learning algorithms or combines deep learning with other AI models such as machine learning and data mining. Islam, Islam & Asraf (2020) introduced a deep learning architecture that combines a CNN and a Long Short-term Memory (LSTM) to identify the COVID-19 from X-ray images automatically. They used the CNN network to extract deep features and LSTM for the detection of the COVID-19 patients. Another study (Islam et al., 2020) introduced architecture to diagnose the COVID-19 using chest X-rays. The architecture combined CNN for feature extraction and recurrent neural network (RNN) for classification to diagnose the COVID-19 from chest X-rays. They used many deep transfer techniques, including VGG19, DenseNet121, InceptionV3, and InceptionResNetV2. They showed that the performance of VGG19-RNN is better than the other compared architectures in terms of accuracy and computational time in our experiments. Aslan et al. (2021) presented a hybrid architecture that combines the CNN and Bi-directional Long Short-term Memories (BiLSTM). They utilized the modified AlexNet (mAlexNet) architecture with chest X-ray images to diagnose the COVID-19. They modified the last three layers of the AlexNet model to build a three-class model classify the COVID-19. The remaining parameters of the original AlexNet model have been preserved. The temporal features obtained from the BiLSTM layer are passed as input to a fully-connected (FC) layer, and the Softmax is used for the classification. Sethy et al. (2020) proposed a hybrid architecture that relies on ImageNet pre-trained models to extract the high-level features and Support Vector Machine (SVM) to detect the COVID-19 cases. Their architecture is a three-class problem to classify the COVID-19 patient from healthy people and pneumonia patients using X-ray images. They showed that the SVM achieved the best results when the features are extracted using the ResNet50 Network. Altan & Karasu (2020) presented a combined architecture to detect the COVID-19 patients from X-ray images. The architecture consists of three different techniques: two-dimensional (2D) curvelet transformation, Chaotic Salp Swarm Algorithm (CSSA) optimization algorithm, and deep learning technique. 2D Curvelet transform is used to get the feature matrix from the patient’s chest X-ray images. An optimization process is done to the feature matrix. The EfficientNetB0 model, based on CNN, is used to classify X-ray images to diagnose the infected COVID-19.

The proposed otld-covid-19 approach

Recently, the COVID-19 pandemic has taken the world’s health care systems by surprise. It will also take a long time to ensure the vaccine’s safety before the general public could use it. As a result, the government’s existing efforts primarily focus on preventing the spread of COVID-19 and forecasting potential pandemic areas. Due to the scarcity of COVID-19 testing kits, particularly in developing countries, alternative diagnosis methods are essential. One solution is to develop COVID-19 diagnosis strategies based on data mining and artificial intelligence. Figure 7 shows the proposed OTLD-COVID-19 approach that consists of five phases: (1) Acquisition phase, (2) Preprocessing phase, (3) Augmentation phase, (4) Training, Classification, and Optimization (TCO) phase, and (5) Deployment phase. The acquisition phase starts with reading the dataset, converting images to the “JPG” format, and resize the images to (64, 64, 3). Normalizing the images X (i.e., ) is performed in the preprocessing phase, followed by noise removal and converting the labels (i.e., classes) from numeric to one-hot encoding. In one-hot encoding, each category is converted into a new column and assigned a 1 or 0 as notation for true/false (e.g. 2 will be [0, 1, 0, 0] for 4 classes).
Figure 7

The proposed OTLD-COVID-19 approach.

The first stage of data augmentation aims to equalize the number of images in each category via shifting, cropping, zooming, and flipping (horizontally, vertically, or both). The TCO phase is the main core of the proposed approach. The first stage splits the augmented dataset into three parts (training, testing, and validation). It applies the transfer learning to obtain a pre-trained CNN model with ImageNet. In this study, twelve different models are tested to create the pre-trained CNN model (i.e., DenseNet121, DenseNet169, DenseNet201, Xception, MobileNet, MobileNetV2, MobileNetV3Small, MobileNetV3Large, EfficientNetB0, ResNet50V2, ResNet101V2, and ResNet152V2). The second data augmentation stage is applied during the training. The training optimizes the learning parameters. The last stage in this phase is the hyperparameters optimization process. The MRFO algorithm is repeated for several iterations (i.e., 15 in the current study) to optimize the main hyperparameters. It defines the population size (N), maximum number of iterations (N) and dataset split ratio (SR). Random positions of the manta-rays are initialized before applying either chain or cyclone foraging according to a random value. MRFO starts the chain foraging if the random value is lower than 0.5. The chain foraging mathematical model is represented in Eqs. (1) and (2) as follows: where is the position of i individual at time t in d dimension, r is a random vector within the range of [0, 1], α is a weight coefficient, and is the plankton with the highest concentration. The cyclone foraging process plays an important part in developing two key driving mechanisms: exploitation and exploration. The exploitation (intensification) aims to find the best candidate solutions in the current search space, called the local search. The exploration (the global search) is concerned with exploring different search space areas to avoid getting stuck in a local minimum. In this foraging process, the best plankton location is used as a reference point, allowing for increased exploitation capabilities by enlarging the fertile regions surrounding the current best solution. Eqs. (3) and (4) mathematically models the exploitation phase. where β is the weight coefficient, T is the maximum number of iterations, and r1 is a random number in the range [0, 1]. Cyclone foraging helps the exploration process by forcing the population members to shift to a random location in the search space. This position is far from their current location as well as the best prey location. This exploration method helps the algorithm extend the global search space and direct the population through the search domain’s unvisited paths. The mathematical model of the exploration process is given by Eqs. (5) and (6) below. where is a random position randomly produced in the search space, and LB and UB are the lower and upper limits of the d dimension, respectively. MRFO shifts between exploration phases according to the ratio between current iteration and the maximum number of iterations (n/N). The exploitation phase is enacted when n/Nrand. After completing either cyclone or chain foraging, summersault foraging takes action. It updates the current position of individuals through the current best solution. The following mathematical formulation (i.e., Eq. (7)) describes the summersault foraging. where S is the somersault factor that decides the somersault range of manta rays and its value equals 2, r2 and r3 are two random numbers in the range [0, 1]. The TCO phase computes the fitness score of each manta-ray and chooses the best individual. The metrics are determined from the evaluation of the trained model on the test part of the dataset. These metrics are used to calculate the overall fitness score after applying the trained model for several epochs (i.e., 64 in the current study). The calculated fitness score is used to update the position of manta-rays. This process repeats until the completion of the iterations (i.e. n = N). Each result is evaluated, and a model is built according to the optimized computed hyperparameters. The model with these optimized hyperparameters is ready to achieve a rigid classification process. The pseudocode of the TCO is represented in Algorithm 1 where the “UpdateMRFO” function that uses the MRFO optimization algorithm is represented in Algorithm 2.
Algorithm 1

The TCO Pesudocode.

Input: Np, Ns, SR, Model // Population Size, Number of Iterations, Dataset Split Ratio, CNN Pre-trained ModelOutput: bestSoFar // The Best CombinationData: X, Y // Images Dataset, Images Labels POs ⇐ The Learning Parameters Optimizers BSs ⇐ The Learning Batch Sizes MLRs ⇐ The Learning Ratios DRs ⇐ The Learning Dropout Ratios Metrics ⇐ The Learning Performance Metrics ns ⇐ 1 // Initialize the Iterations Counter fitnessScores ⇐ [ ] trainX, validationX, testX, trainY, validationY, testYDatasetSplit(X, Y, SR)// Get Np Solutions using a Random Initialization Process populationInitializePopulation(POs, BSs, MLRs, DRs, Np)// Find the Population Best Combinations using the Iterative Optimization Approach while nsNs do// Find the Population Fitness Score Values of the Current Population np ⇐ 1 // Initialize the Populations Counter while npNp do solutionGetSolution(population, np) // Get the Solution at Index of np ModelLearnCNN(Model, solution, Metrics, trainX, trainY, validationX, validationY)// Apply the CNN Model Learning Proce (i.e. Train and Validate) metricsTestCNN(Model, Metrics, testX, testY) // The the Optimized CNN using the Required Metrics fitnessScoreGetOverallFitnessScore(metrics) // Get the Overall Fitness Score using the MetricsResults fitnessScores.push(fitnessScore) // Append the Latest Fitness Score to the End of the List npnp + 1 // Increment the Populations Counter end sortedPopulationSortFitnessScores(population, fitnessScores) // Sort the Population concrening the Fitness Score Values populationUpdateMRFO(sortedPopulation, POs, BSs, MLRs, DRs, Np) // Update the Population using the MRFO Algorithm nsns + 1 // Increment the Iterations Counter end bestSoFarGetTop(population) // Extract the Overall Top Combination
Algorithm 2

The Manta Ray Foragin Optimization (MRFO) Algorithm Pesudocode.

Input: population, bestFitness, Ns, ns, LB, UB, D // Population, The Best Fitness Score, Number of Iterations, Current Iteration, Lower Bound, Upper Bound, Number of DimensionsOutput: newPopulation // The New PopulationnewPopulationCopy(population) Create a Copy from the Current Sorted Population bestSolutionnewPopulation [1] Get the Best Solution (i.e., First Solution) From the Sorted PopulationcoefnsNSnp ⇐ 1 Initialize the Populations Counterwhile (npNp) dorrandom() // A Random Number in the Range [0, 1] r1 ⇐ random() // A Random Number in the Range [0, 1] α ⇐ 2.0 × r × ||logr|| // Calculate the Value of AlphafactorNsns+1.0NS×1.0β ⇐ 2.0 × exp(r1 × factor) × sin(2.0 × π × r1) // Calculate the Value of Betaif (random() < 0.5) then   // Cyclone Foraging   if (coef < random()) then    randLB + random(D) × (UBLB)   if (np = 1) then     newPopulation[np] ⇐ rand + r × (randnewPopulation[np]) + β × (randnewPopulation[np])   end   else    newPopulation[np] ⇐ rand + r × (newPopulation[np − 1] − newPopulation[np]) + β × (randnewPopulation[np])   end  end  else   if (np = 1) then     newPopulation[np] ⇐ bestSolution + r × (bestSolutionnewPopulation[np]) + β × (bestSolutionnewPopulation[np])   end   else    newPopulation[np] ⇐ bestSolution + r × (newPopulation[np − 1] − newPopulation[np]) + β × (bestSolutionnewPopulation[np])   end  endendelse  // Chain Foraging  if np = 1 then    newPopulation[np] ⇐ newPopulation[np] + r × (bestSolutionnewPopulation[np]) + α × (bestSolutionnewPopulation[np])  end  else   newPopulation[np] ⇐ newPopulation[np] + r × (newPopulation[np − 1] − newPopulation[np]) + α × (bestSolutionnewPopulation[np])  endendfitnessScoreGetFitnessScore(newPopulation[np]) // Get the Overall Fitness Score After Re-Learning the Model if (fitnessScore > bestFitness) then  bestSolutionnewPopulation[np]  bestFitnessfitnessScoreend Somersault Foraging S ⇐ 2.0r2 ⇐ random()r3 ⇐ random()newPopulation[np] ⇐ newPopulation[np] + S × (r2 × bestSolutionr3 × newPopulation[np])fitnessScoreGetFitnessScore(newPopulation[np]) // Get the Overall Fitness Score After Re-Learning the Model if (fitnessScore > bestFitness) then   bestSolutionnewPopulation[np]  bestFitnessfitnessScoreendend
The deployment phase of the proposed approach uses computed hyperparameters to build a classification model. Each model evaluates the COVID-19 dataset to classify the cases into the main four categories (i.e., “Normal,” “Viral Pneumonia,” “COVID-19”, and “Bacterial Pneumonia”). The next section will discuss the experimental results of the proposed OTLD-COVID-19 approach compared to recent state-of-the-art approaches and explain its effectiveness.

Experiments, results and discussion

This section presents the different applied experiments with the corresponding results and discussions. It also presents the used dataset in the current study and ends by applying a comparative study between the current study and a set of state-of-the-art studies.

Dataset and experiments configurations

The adopted datasets in this study are real datasets used to distinguish COVID-19 from common pneumonia types. The proposed dataset is unified and collected from eight different public data sources described in Table 3 and graphically illustrated in Fig. 8. The dataset consists of chest X-ray images in four classes: radiographs of normal cases, viral pneumonia, COVID-19 pneumonia, and bacterial pneumonia. The total number of cases in the collected dataset is 12,933. Table 4 summarizes the common experiments configurations.
Table 3

The used datasets description.

#DatasetCOVID-19P-virusP-bacteriaNormal
1COVID-19 Radiography Database (Cohen et al., 2020).40838460
2Pneumonia (virus) Vs. COVID-19 (Mahasin, 2020).701,49300
3COVID-19 X-ray images using CNN (Srikar, 2020).14000144
4COVID-19 X-ray Images (Das, 2020).88001,002
5COVID-19 Patients Lungs X Ray Images 10000 (Sajid, 2020).700028
6COVID-19 Chest X Rays (Sreeraman, 2020).690079
7COVID-19 Dataset (Riaz, 2020).250025
8Curated Chest X-Ray Image Dataset for COVID-19 (Sait, 2020).1,2811,6563,0013,270
Total2,1513,1873,0474,548
Figure 8

Distribution of the datasets and their categories.

Table 4

Experiments configurations summarization.

KeyValue
Dataset8 Resources
Categories“Normal”, “P-Viral”, “P-Bacterial”, and “COVID-19”
Dataset Size12,933
Pre-trained ModelsDenseNet121, DenseNet169, DenseNet201, Xception, MobileNet, MobileNetV2, MobileNetV3Small, MobileNetV3Large, EfficientNetB0, ResNet50V2, ResNet101V2, and ResNet152V2
Parameters InitializerImageNet
Parameters OptimizersAdam, NAdam, Ftrl, AdaDelta, AdaGrad, AdaMax, RMSProp, and SGD
Output Activation FunctionSoftMax
Model Learn Ratios[0 : 5 : 100]%
Batch Sizes[8 : 8 : 100]
Dropout Ratios[1 : 1 : 60]%
Number of Epochs8 Epochs
Performance MetricsAccuracy, Loss, Precision, F1-score, AUC, Dice Coef., IoU Coef., Specificity, and Recall
Number of Iterations Ns20 Iterations
Population Size Np10 Candidates
Split Ratio SR85% to 15%
Image Size(64, 64, 3)
Data AugmentationApplied (in Two Stages)
Training EnvironmentGoogle Colab (using its GPU)

Performance metrics

During the next experiments, there are different metrics to evaluate the performance of the “OTLD-COVID-19” approach. At first, the confusion matrix that represents a summary of predicted results is constructed. The confusion matrix has four values as follows: True Positive (TP) occurs when the actual class of the data is positive (True) and the predicted is also positive (True). True Negative (TN) occurs when the actual class of the data is negative (False), and the predicted is also negative (False). False Positive (FP) occurs when the actual class of the data is negative (False) while the predicted is positive (Tues). False Negative (FN) occurs when the actual class of the data is positive (True), and the predicted is negative (False). Different formulas are used as a summarization of the confusion matrix. Table 5 depicts several performance metrics, including Accuracy, Recall, Precision, F1-score, and Loss. Among these metrics, accuracy has the most attention for the results of deep learning classifiers in the condition that the data is well balanced and not skewed for a specific class. It is the fraction of predictions the model classified correctly to all the predictions of the model. Precision is used as an evaluation metric to ensure our prediction. Recall or Sensitivity (True Positive Rate) is essential to understand how complete the results are. F1-score is an overall measure of the model’s accuracy that combines precision and recall. Specificity (False Positive Rate) is the ratio between the false-negative data that is mistakenly considered positive and all negative data. AUC is defined as the area under the Receiver Operating Characteristic (ROC) Curve.
Table 5

The used performance metrics.

MetricDefinitionFormula
AccuracyThe ratio between the correct predictions made by the model and all kinds’ predictions made.TP+TNTP+TN+FP+FN
PrecisionThe ratio between the true positive predicted values and full positive predicted values.TPTP+FP
Recall or SensitivityThe ratio between the true positive values of prediction and all predicted values.TPTP+FN
F1-scoreTwice the ratio between the multiplication to the summation of precision and recall metrics.2×TP2×TP+FP+FN
SpecificityThe ratio between the false-negative data that is mistakenly considered positive and all negative data.FPTN+FP
AUCPlotting the cumulative distribution function of the True Positive Rate (TPR) in the y-axis versus the cumulative distribution function of the False Positive Rate (FPR) on the X-axis.TPR=TPTP+FN and FPR=FPFP+TN
IoU CoefficientThe ratio between the area of intersection and area of union.TPTP+TN+FN
Dice CoefficientTwice the ratio between the true positive predicted values and all other values.2×TPTP+TN+FP+FN
LossThe distance between the true values of the problem and the values predicted by the model.l(y,p)=Σc=1Myo,c×logpo,c
The ROC curve is generated by plotting the True Positive (TP) cumulative distribution function in the y-axis versus the False Positive (FP) cumulative distribution function on the X-axis. The AUC is a single-valued metric, and as the AUC value gets higher as the performance of the model increases and easily distinguishes between the different classes. IoU (Intersection over Union) score is considered good when its value is greater than 0.5. The loss is the distance between the true values of the problem and the values predicted by the model. The lower the loss, the better a model unless the model has overfitted to the training data. In the loss formula, M is the number of classes, l is the loss value, and p is the predicted value.

Results and Discussion

DenseNet121: Table 6 reports the DenseNet121 top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 0.1268, 97.39%, 0.9738, 97.44%, 97.33%, 99.15%, 0.9962, 0.9733, 0.9750, and 0.9783 respectively while the top-1 results, concerning the accuracy, were 0.1268, 97.39%, 0.9738, 97.44%, 97.33%, 99.15%, 0.9962, 0.9733, 0.9750, and 0.9783 respectively. They were reported by the AdaMax parameters optimizer, batch size value of 40, dropout ratio of 0.075, and model learning ratio of 55%. Table 7 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 6

DenseNet121 learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1RMSProp320.60010%0.261489.59%0.896389.96%89.31%96.68%0.98660.89310.89530.9110
2AdaGrad320.02525%0.277190.36%0.904490.69%90.20%96.91%0.98540.90200.91410.9249
3AdaMax320.60010%0.252990.71%0.905790.93%90.22%97.00%0.98750.90220.89960.9149
4AdaMax240.55010%0.239290.82%0.906691.12%90.22%97.07%0.98910.90220.89770.9143
5NAdam240.02515%0.239291.56%0.916091.78%91.43%97.27%0.98850.91430.91110.9245
6AdaMax480.32520%0.281091.69%0.917791.99%91.56%97.34%0.98470.91560.92580.9345
7AdaMax720.07550%0.229291.99%0.920892.29%91.88%97.44%0.98950.91880.91290.9263
8Adam160.00015%0.210892.64%0.926892.92%92.45%97.65%0.99090.92450.92250.9344
9AdaMax320.55010%0.200292.69%0.927693.21%92.32%97.76%0.99170.92320.91080.9260
10AdaMax320.55010%0.194192.76%0.927292.96%92.49%97.67%0.99270.92490.91290.9277
11Adam240.00020%0.189793.03%0.929793.42%92.54%97.83%0.99280.92540.90930.9255
12NAdam240.02515%0.233493.18%0.932193.34%93.09%97.79%0.98910.93090.94200.9487
13Adam640.02555%0.184694.61%0.946094.76%94.45%98.26%0.99200.94450.95000.9562
14AdaMax400.07555%0.149496.04%0.960496.08%96.01%98.69%0.99450.96010.96660.9701
15AdaMax400.07555%0.126897.39%0.973897.44%97.33%99.15%0.99620.97330.97500.9783

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 7

DenseNet121 model correlation results.

MetricBatch sizeDropout ratioModel learn ratio
Loss−0.10530.2908−0.5941
Accuracy0.2305−0.46260.7213
F10.2410−0.47630.7279
Precision0.2282−0.45800.7132
Recall0.2517−0.49210.7397
Specificity0.2251−0.45260.7095
AUC0.0421−0.24920.5048
Sensitivity0.2517−0.49210.7397
IoU Coef0.2899−0.55230.7699
Dice Coef0.2781−0.54390.7694
Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union. DenseNet169: Table 8 reports the DenseNet169 top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 0.0523, 98.47%, 0.9849, 98.50%, 98.47%, 99.50%, 0.9983, 0.9847, 0.9860, and 0.9879 respectively while the top-1 results, concerning the accuracy, were 0.0523, 98.47%, 0.9849, 98.50%, 98.47%, 99.50%, 0.9983, 0.9847, 0.9860, and 0.9879 respectively. They were reported by the SGD parameters optimizer, batch size value of 88, dropout ratio of 0.025, and model learning ratio of 90%. Table 9 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 8

DenseNet169 learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1AdaGrad960.000100%0.152494.94%0.949195.03%94.79%98.35%0.99430.94790.95170.9580
2AdaMax560.00060%0.186494.95%0.949395.04%94.82%98.35%0.99250.94820.95270.9585
3SGD880.02590%0.134995.09%0.951195.18%95.03%98.40%0.99510.95030.95380.9604
4NAdam720.00070%0.141095.28%0.953095.40%95.20%98.47%0.99480.95200.95500.9613
5AdaMax960.000100%0.185495.72%0.956995.71%95.67%98.57%0.99200.95670.96380.9676
6AdaMax560.00055%0.165995.85%0.959496.01%95.86%98.67%0.99400.95860.96430.9684
7NAdam720.00070%0.126695.91%0.958795.97%95.77%98.66%0.99610.95770.95680.9635
8NAdam960.00090%0.105296.05%0.960596.11%95.99%98.70%0.99740.95990.95610.9634
9RMSProp640.02570%0.169896.13%0.960996.13%96.06%98.71%0.99430.96060.96670.9704
10NAdam880.00085%0.112096.65%0.966796.72%96.62%98.91%0.99610.96620.97180.9751
11NAdam880.00085%0.097596.85%0.968896.97%96.80%98.99%0.99620.96800.96960.9737
12NAdam560.00055%0.085497.36%0.973597.40%97.31%99.13%0.99680.97310.97340.9774
13AdaMax640.00065%0.082398.05%0.980698.07%98.04%99.36%0.99690.98040.98110.9835
14SGD880.02590%0.061798.38%0.983898.40%98.36%99.47%0.99810.98360.98200.9847
15SGD880.02590%0.052398.47%0.984998.50%98.47%99.50%0.99830.98470.98600.9879

Note:

POO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 9

DenseNet169 model correlation results.

MetricBatch sizeDropout ratioModel learn ratio
Loss−0.2079−0.2777−0.1250
Accuracy0.01050.3294−0.0119
F10.00680.3264−0.0180
Precision0.00370.3202−0.0227
Recall0.00940.3315−0.0139
Specificity0.00500.3222−0.0209
AUC0.19630.30470.0951
Sensitivity0.00940.3315−0.0139
IoU Coef−0.03080.3537−0.0244
Dice Coef−0.01750.3557−0.0202
Note: POO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union. DenseNet201: Table 10 reports the DenseNet201 top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 0.1104, 96.69%, 0.9675, 96.90%, 96.61%, 98.97%, 0.9967, 0.9661, 0.9667, and 0.9711 respectively while the top-1 results, concerning the accuracy, were 0.1153, 96.69%, 0.9675, 96.90%, 96.61%, 98.97%, 0.9958, 0.9661, 0.9667, and 0.9711 respectively. They were reported by the SGD parameters optimizer, batch size value of 80, dropout ratio of 0.000, and model learning ratio of 45%. Table 11 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 10

DenseNet201 learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1SGD720.00035%0.226892.58%0.926892.90%92.46%97.65%0.98850.92460.92750.9374
2RMSProp560.05050%2.037092.60%0.925692.63%92.49%97.55%0.98120.92490.93900.9451
3AdaMax800.12560%0.225993.25%0.932693.37%93.15%97.80%0.98920.93150.93720.9448
4SGD800.00045%0.191293.50%0.934493.55%93.32%97.86%0.99250.93320.94150.9488
5AdaMax560.00025%0.183393.59%0.936193.76%93.46%97.93%0.99330.93460.93610.9452
6SGD880.00040%0.187493.80%0.938393.91%93.76%97.97%0.99280.93760.94560.9521
7SGD800.00040%0.205993.93%0.939794.17%93.78%98.06%0.98980.93780.94110.9485
8AdaMax560.00025%0.177694.10%0.941794.38%93.96%98.13%0.99310.93960.94290.9508
9RMSProp800.00040%0.244994.99%0.949795.00%94.95%98.33%0.98860.94950.96160.9646
10AdaMax560.00025%0.166095.02%0.950495.15%94.93%98.39%0.99290.94930.95260.9586
11SGD880.00040%0.155195.15%0.951695.24%95.07%98.42%0.99440.95070.95300.9590
12SGD880.00050%0.119396.09%0.960896.23%95.94%98.75%0.99600.95940.96030.9660
13SGD800.00040%0.110496.14%0.961596.28%96.02%98.76%0.99670.96020.95950.9658
14RMSProp720.00040%0.135696.27%0.963096.37%96.24%98.79%0.99520.96240.96530.9696
15SGD800.00045%0.115396.69%0.967596.90%96.61%98.97%0.99580.96610.96670.9711

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 11

DenseNet201 model correlation results.

MetricBatch sizeDropout ratioModel learn ratio
Loss−0.41540.33970.2799
Accuracy0.3403−0.3941−0.0038
F10.3359−0.3995−0.0169
Precision0.3299−0.4090−0.0302
Recall0.3415−0.3895−0.0032
Specificity0.3294−0.4085−0.0301
AUC0.3475−0.4665−0.2434
Sensitivity0.3415−0.3895−0.0032
IoU Coef0.3457−0.33460.1004
Dice Coef0.3447−0.36070.0696
Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union. Xception: Table 12 reports the Xception top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 0.0861, 97.78%, 0.9783, 97.90%, 97.76%, 99.30%, 0.9979, 0.9776, 0.9750, and 0.9789 respectively while the top-1 results, concerning the accuracy, were 0.0888, 97.78%, 0.9783, 97.90%, 97.76%, 99.30%, 0.9961, 0.9776, 0.9750, and 0.9789 respectively. They were reported by the RMSProp parameters optimizer, batch size value of 80, dropout ratio of 0.525, and model learning ratio of 70%. Table 13 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 12

Xception learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1RMSProp800.52570%0.157495.12%0.951695.28%95.03%98.43%0.99330.95030.95600.9617
2NAdam240.10015%0.175795.26%0.953095.37%95.23%98.46%0.99250.95230.95940.9634
3RMSProp960.60075%0.290195.50%0.954995.51%95.47%98.50%0.98940.95470.96380.9669
4RMSProp880.55070%0.157195.69%0.957195.84%95.58%98.62%0.99460.95580.96300.9671
5NAdam240.12520%0.127995.83%0.957295.82%95.61%98.61%0.99610.95610.95360.9610
6AdaMax560.37555%0.139495.89%0.958295.95%95.70%98.65%0.99430.95700.94870.9573
7NAdam240.15025%0.111596.06%0.960696.11%96.01%98.71%0.99710.96010.95650.9636
8NAdam240.10015%0.113896.13%0.961596.30%96.00%98.77%0.99690.96000.95450.9620
9NAdam240.10015%0.109796.17%0.961696.22%96.10%98.74%0.99680.96100.95580.9631
10NAdam240.10015%0.095796.31%0.962896.37%96.19%98.79%0.99790.96190.95730.9647
11NAdam240.10015%0.103196.41%0.964096.48%96.32%98.83%0.99750.96320.95670.9641
12RMSProp960.60075%0.137496.79%0.967696.78%96.74%98.93%0.99400.96740.97280.9756
13NAdam240.10015%0.086196.99%0.969997.07%96.91%99.03%0.99790.96910.96270.9692
14RMSProp880.55070%0.161897.12%0.971097.15%97.05%99.05%0.99670.97050.97230.9757
15RMSProp800.52570%0.088897.78%0.978397.90%97.76%99.30%0.99610.97760.97500.9789

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 13

Xception model correlation results.

MetricBatch sizeDropout ratioModel learn ratio
Loss0.54630.52750.5068
Accuracy0.10730.10060.0943
F10.11690.10940.1023
Precision0.10420.09790.0923
Recall0.12880.12020.1117
Specificity0.10050.09430.0888
AUC−0.5791−0.5741−0.5661
Sensitivity0.12880.12020.1117
IoU Coef0.63940.60980.5760
Dice Coef0.55840.53240.5025
Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union. MobileNet: Table 14 reports the MobileNet top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 0.1247, 96.40%, 0.9636, 96.39%, 96.33%, 98.80%, 0.9952, 0.9633, 0.9664, and 0.9706 respectively while the top-1 results, concerning the accuracy, were 0.1247, 96.40%, 0.9636, 96.39%, 96.33%, 98.80%, 0.9952, 0.9633, 0.9664, and 0.9706 respectively. They were reported by the NAdam parameters optimizer, batch size value of 56, dropout ratio of 0.050, and model learning ratio of 25%. Table 15 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 14

MobileNet learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1SGD960.325100%0.353887.74%0.877988.32%87.29%96.15%0.97900.87290.88640.9013
2Adam80.20030%0.298187.75%0.878288.77%86.91%96.33%0.98240.86910.87200.8921
3NAdam400.05020%0.339389.47%0.894889.64%89.33%96.56%0.98060.89330.91210.9215
4SGD960.325100%0.301789.97%0.900290.41%89.64%96.83%0.98230.89640.90310.9163
5NAdam240.10015%0.259390.09%0.901190.59%89.66%96.90%0.98700.89660.89450.9105
6Adam320.02515%0.277591.44%0.913891.63%91.13%97.22%0.98640.91130.91770.9287
7NAdam480.05025%0.255392.34%0.923092.60%92.01%97.55%0.98610.92010.92970.9383
8AdaMax240.45045%0.235192.39%0.923992.63%92.16%97.56%0.98900.92160.92380.9345
9RMSProp240.47545%0.255292.78%0.928293.04%92.62%97.69%0.98780.92620.92890.9387
10NAdam720.07535%0.222693.43%0.934293.56%93.28%97.86%0.98850.93280.93740.9456
11SGD320.60060%0.228193.68%0.936193.67%93.55%97.89%0.98960.93550.94530.9513
12Adam80.17530%0.177794.35%0.942394.63%93.84%98.23%0.99340.93840.91870.9334
13SGD160.60050%0.174094.51%0.944694.66%94.26%98.23%0.99290.94260.93000.9418
14AdaMax240.45045%0.144495.85%0.957696.06%95.46%98.69%0.99380.95460.94830.9567
15NAdam560.05025%0.124796.40%0.963696.39%96.33%98.80%0.99520.96330.96640.9706

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 15

MobileNet model correlation results.

MetricBatch sizeDropout ratioModel learn ratio
Loss0.3770−0.19140.2906
Accuracy−0.27030.2201−0.2438
F1−0.26660.2226−0.2407
Precision−0.29380.2259−0.2452
Recall−0.24240.2191−0.2363
Specificity−0.29900.2258−0.2463
AUC−0.45680.2247−0.3325
Sensitivity−0.24240.2191−0.2363
IoU Coef−0.05510.1504−0.1983
Dice Coef−0.09810.1654−0.2108
Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union. MobileNetV2: Table 16 reports the MobileNetV2 top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 0.6045, 81.54%, 0.8158, 81.69%, 81.48%, 93.91%, 0.9417, 0.8148, 0.8635, and 0.8728 respectively while the top-1 results, concerning the accuracy, were 2.5140, 81.54%, 0.8158, 81.69%, 81.48%, 93.91%, 0.9352, 0.8148, 0.8635, and 0.8728 respectively. They were reported by the RMSProp parameters optimizer, batch size value of 16, dropout ratio of 0.425, and model learning ratio of 55%. Table 17 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 16

MobileNetV2 learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1Adam80.0000%0.619875.70%0.752478.53%72.31%93.42%0.93760.72310.76590.7967
2Adam80.0000%0.619875.70%0.752478.53%72.31%93.42%0.93760.72310.76590.7967
3Adam80.0000%0.611075.87%0.750578.32%72.13%93.34%0.93990.72130.77220.8016
4Adam80.0000%0.611075.87%0.750578.32%72.13%93.34%0.93990.72130.77220.8016
5Adam80.0000%0.629076.37%0.759579.11%73.14%93.57%0.93700.73140.77230.8015
6Adam80.0000%0.629076.37%0.759579.11%73.14%93.57%0.93700.73140.77230.8015
7Adam80.0000%0.610076.69%0.761778.93%73.67%93.44%0.93980.73670.77740.8058
8Adam80.0000%0.606876.70%0.762579.13%73.66%93.52%0.93990.73660.77280.8031
9Adam80.0000%0.604577.02%0.765679.36%74.03%93.58%0.94170.74030.77790.8071
10Adam80.0000%0.604577.02%0.765679.36%74.03%93.58%0.94170.74030.77790.8071
11Adam80.0000%0.604577.02%0.765679.36%74.03%93.58%0.94170.74030.77790.8071
12Adam80.0000%0.604577.02%0.765679.36%74.03%93.58%0.94170.74030.77790.8071
13Adam80.0000%0.604577.02%0.765679.36%74.03%93.58%0.94170.74030.77790.8071
14Adam80.0000%0.604577.02%0.765679.36%74.03%93.58%0.94170.74030.77790.8071
15RMSProp160.42555%2.514081.54%0.815881.69%81.48%93.91%0.93520.81480.86350.8728

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 17

MobileNetV2 model correlation results.

MetricBatch SizeDropout RatioModel Learn Ratio
Loss0.99980.99980.9998
Accuracy0.92770.92770.9277
F10.92410.92410.9241
Precision0.86820.86820.8682
Recall0.93970.93970.9397
Specificity0.75330.75330.7533
AUC−0.5546−0.5546−0.5546
Sensitivity0.93970.93970.9397
IoU Coef0.98380.98380.9838
Dice Coef0.97930.97930.9793
Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union. MobileNetV3Small: Table 18 reports the MobileNetV3Small top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 0.7490, 80.06%, 0.7995, 80.50%, 79.43%, 97.76%, 0.9380, 0.7943, 0.8183, and 0.8381 respectively while the top-1 results, concerning the accuracy, were 27.860, 80.06%, 0.7995, 80.50%, 79.43%, 93.59%, 0.9270, 0.7943, 0.8159, and 0.8365 respectively. They were reported by the RMSProp parameters optimizer, batch size value of 8, dropout ratio of 0.075, and model learning ratio of 100%. Table 19 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 18

MobileNetV3Small learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1Adam80.0000%0.989455.61%0.498371.14%38.65%94.75%0.83180.38650.56490.6192
2Adam80.0000%0.989455.61%0.498371.14%38.65%94.75%0.83180.38650.56490.6192
3AdaDelta400.07555%0.987758.12%0.562763.96%50.45%90.53%0.83700.50450.63320.6749
4Adam80.0000%1.032058.49%0.318075.81%20.50%97.76%0.82470.20500.53570.5922
5Adam80.0000%1.032058.49%0.318075.81%20.50%97.76%0.82470.20500.53570.5922
6Adam80.0000%1.032058.49%0.318075.81%20.50%97.76%0.82470.20500.53570.5922
7NAdam80.00025%0.917760.73%0.594864.47%55.34%89.82%0.86470.55340.65700.6989
8Adam80.0000%0.988161.23%0.441670.65%32.55%95.50%0.84160.32550.56130.6159
9Adam80.0000%0.988161.23%0.441670.65%32.55%95.50%0.84160.32550.56130.6159
10Adam80.0000%0.973662.57%0.434173.52%31.22%96.25%0.84920.31220.55680.6131
11Adam80.0000%0.973662.57%0.434173.52%31.22%96.25%0.84920.31220.55680.6131
12AdaDelta400.07555%0.904163.23%0.616268.65%56.10%91.44%0.86360.56100.66050.7007
13AdaGrad80.02540%0.749070.61%0.671478.95%58.66%94.81%0.90460.58660.67090.7157
14RMSProp80.075100%1.121079.15%0.789779.65%78.32%93.33%0.93800.78320.81830.8381
15RMSProp80.075100%27.86080.06%0.799580.50%79.43%93.59%0.92700.79430.81590.8365

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 19

MobileNetV3Small model correlation results.

MetricBatch sizeDropout ratioModel learn ratio
Loss−0.10660.43540.5648
Accuracy−0.12790.63680.8362
F10.19180.76710.8984
Precision−0.53840.09360.2804
Recall0.21590.79670.9284
Specificity−0.5904−0.6059−0.5832
AUC−0.07280.65550.8645
Sensitivity0.21590.79670.9284
IoU Coef0.13550.81060.9677
Dice Coef0.12480.80310.9639
Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union. MobileNetV3Large: Table 20 reports the MobileNetV3Large top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 0.5192, 79.57%, 0.7851, 83.15%, 74.50%, 95.12%, 0.9531, 0.7450, 0.7550, and 0.7931 respectively while the top-1 results, concerning the accuracy, were 0.5192, 79.57%, 0.7851, 83.15%, 74.50%, 94.97%, 0.9531, 0.7450, 0.7550, and 0.7931 respectively. They were reported by the AdaMax parameters optimizer, batch size value of 16, dropout ratio of 0.175, and model learning ratio of 20%. Table 21 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 20

MobileNetV3Large learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1Adam80.0000%0.822668.66%0.644877.23%55.62%94.52%0.88660.55620.64150.6907
2Adam80.0000%0.822668.66%0.644877.23%55.62%94.52%0.88660.55620.64150.6907
3Adam80.0000%0.789870.73%0.660179.48%56.74%95.12%0.89710.56740.64570.6953
4Adam80.0000%0.789870.73%0.660179.48%56.74%95.12%0.89710.56740.64570.6953
5Adam80.0000%0.789870.73%0.660179.48%56.74%95.12%0.89710.56740.64570.6953
6Adam80.0000%0.804271.07%0.651777.69%56.44%94.60%0.89270.56440.63520.6870
7Adam80.0000%0.804271.07%0.651777.69%56.44%94.60%0.89270.56440.63520.6870
8Adam80.0000%0.804271.07%0.651777.69%56.44%94.60%0.89270.56440.63520.6870
9Adam80.0000%0.804271.07%0.651777.69%56.44%94.60%0.89270.56440.63520.6870
10AdaDelta80.12515%0.671371.79%0.705676.87%65.40%93.45%0.92120.65400.69680.7403
11AdaDelta80.12515%0.663271.90%0.706577.63%65.00%93.76%0.92300.65000.70590.7474
12AdaDelta80.10015%0.711371.99%0.702877.77%64.30%93.87%0.91170.64300.69550.7374
13AdaDelta80.12515%0.672572.91%0.720478.43%66.77%93.88%0.92140.66770.71660.7553
14AdaDelta80.12515%0.671173.09%0.720278.62%66.62%93.97%0.92170.66620.71630.7552
15AdaMax160.17520%0.519279.57%0.785183.15%74.50%94.97%0.95310.74500.75500.7931

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 21

MobileNetV3Large model correlation results.

MetricBatch sizeDropout ratioModel learn ratio
Loss−0.7073−0.9595−0.9387
Accuracy0.87010.75880.7283
F10.70910.95430.9381
Precision0.83930.34400.3038
Recall0.64940.98040.9681
Specificity0.2680−0.6046−0.6364
AUC0.69540.95390.9323
Sensitivity0.64940.98040.9681
IoU Coef0.58880.98430.9745
Dice Coef0.61020.98360.9723
Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union. EfficientNetB0: Table 22 reports the EfficientNetB0 top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 1.0970, 51.23%, 0.5070, 92.71%, 48.72%, 100.0%, 0.8072, 0.4872, 0.6478, and 0.6724 respectively while the top-1 results, concerning the accuracy, were 1.3980, 51.23%, 0.5070, 52.92%, 48.72%, 85.55%, 0.8003, 0.4872, 0.6478, and 0.6724 respectively. They were reported by the RMSProp parameters optimizer, batch size value of 64, dropout ratio of 0.375, and model learning ratio of 80%. Table 23 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 22

EfficientNetB0 learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1Ftrl960.600100%1.386027.61%0.00000.00%0.00 %100.00%0.51740.00000.45460.5001
2NAdam240.10015%1.403027.61%0.00000.00%0.00 %100.00%0.52630.00000.45780.5029
3Adam80.0255%1.385027.61%0.00000.00%0.00 %100.00%0.52830.00000.45510.5007
4AdaMax160.25020%1.416027.61%0.112563.48%6.30%98.86%0.54790.06300.47190.5173
5SGD880.275100%1.827027.77%0.286833.40%25.25%83.21%0.62040.25250.53410.5557
6AdaMax640.40065%1.280027.97%0.294633.80%26.22%82.82%0.71800.26220.51460.5603
7SGD880.37595%1.391028.90%0.023025.92%1.22%99.34%0.55430.01220.46250.5081
8AdaDelta480.15050%1.388031.70%0.216339.42%15.12%92.22%0.60850.15120.48910.5326
9AdaGrad480.05050%2.713036.25%0.361536.23%36.07%78.83%0.69230.36070.56420.5772
10AdaMax160.27520%1.257036.43%0.197192.71%11.26%99.81%0.67120.11260.48650.5354
11AdaGrad880.40060%1.210038.65%0.322641.50%26.59%87.51%0.73170.26590.50880.5591
12AdaGrad480.15080%1.159045.03%0.384951.30%31.05%90.15%0.76430.31050.55740.6049
13AdaMax240.40030%4.417045.29%0.433545.38%41.55%83.31%0.59670.41550.51440.5513
14SGD880.425100%1.097050.95%0.490454.48%44.71%87.54%0.80720.44710.61180.6525
15RMSProp640.37580%1.398051.23%0.507052.92%48.72%85.55%0.80030.48720.64780.6724

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 23

EfficientNetB0 model correlation results.

MetricBatch sizeDropout ratioModel learn ratio
Loss−0.2555−0.0024−0.2248
Accuracy0.11050.19890.2437
F10.20520.16620.3176
Precision−0.16740.0925−0.0458
Recall0.22640.17310.3315
Specificity−0.2941−0.0617−0.3210
AUC0.30960.14850.3918
Sensitivity0.22640.17310.3315
IoU Coef0.29580.11160.4449
Dice Coef0.29690.16300.4439
Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union. ResNet50V2: Table 24 reports the ResNet50V2 top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 0.0792, 97.46%, 0.9737, 97.60%, 97.14%, 99.20%, 0.9984, 0.9714, 0.9721, and 0.9750 respectively while the top-1 results, concerning the accuracy, were 0.0792, 97.46%, 0.9737, 97.60%, 97.14%, 99.20%, 0.9984, 0.9714, 0.9635, and 0.9703 respectively. They were reported by the AdaGrad parameters optimizer, batch size value of 16, dropout ratio of 0.100, and model learning ratio of 30%. Table 25 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 24

ResNet50V2 learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1RMSProp800.50025%0.338289.52%0.894189.78%89.06%96.62%0.97970.89060.89690.9114
2AdaMax880.47510%0.376789.53%0.895989.89%89.30%96.65%0.97790.89300.90770.9180
3AdaGrad160.10030%0.212491.81%0.918892.36%91.43%97.48%0.99110.91430.90550.9213
4NAdam80.05015%0.214491.84%0.920292.39%91.66%97.48%0.99030.91660.90660.9222
5AdaGrad160.12535%0.232692.09%0.921792.42%91.93%97.49%0.98910.91930.92660.9362
6NAdam80.05015%0.214492.96%0.930293.13%92.92%97.71%0.99060.92920.93350.9423
7NAdam80.05015%0.184393.33%0.932793.56%92.99%97.87%0.99330.92990.93040.9407
8NAdam80.05015%0.211993.56%0.935493.62%93.46%97.88%0.99050.93460.93450.9436
9AdaMax720.45025%0.226694.23%0.942794.38%94.17%98.13%0.98850.94170.95130.9565
10AdaGrad160.10030%0.171794.35%0.943494.74%93.96%98.26%0.99370.93960.93040.9419
11AdaMax720.45025%0.116196.42%0.964196.50%96.32%98.84%0.99640.96320.96320.9684
12AdaGrad160.10030%0.120496.83%0.967896.87%96.69%98.96%0.99560.96690.97210.9750
13AdaGrad160.10030%0.086896.96%0.970097.14%96.87%99.05%0.99780.96870.96370.9700
14AdaGrad160.10030%0.092097.13%0.971697.31%97.01%99.11%0.99770.97010.96000.9671
15AdaGrad160.10030%0.079297.46%0.973797.60%97.14%99.20%0.99840.97140.96350.9703

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 25

ResNet50V2 model correlation results.

MetricBatch sizeDropout ratioModel learn ratio
Loss0.54480.5166−0.5133
Accuracy−0.3539−0.33190.4854
F1−0.3611−0.33950.4834
Precision−0.3719−0.35030.4991
Recall−0.3509−0.32910.4691
Specificity−0.3721−0.35050.5021
AUC−0.6208−0.59350.4764
Sensitivity−0.3509−0.32910.4691
IoU Coef−0.1844−0.16420.4113
Dice Coef−0.2259−0.20430.4341
Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union. ResNet101V2: Table 26 reports the ResNet101V2 top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 0.0673, 98.47%, 0.9845, 98.45%, 98.45%, 99.48%, 0.9984, 0.9845, 0.9862, and 0.9878 respectively while the top-1 results, concerning the accuracy, were 0.0673, 98.47%, 0.9845, 98.45%, 98.45%, 99.48%, 0.9972, 0.9845, 0.9862, and 0.9878 respectively. They were reported by the Adam parameters optimizer, batch size value of 56, dropout ratio of 0.075, and model learning ratio of 30%. Table 27 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 26

ResNet101V2 learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1Adam320.02515%0.121496.23%0.961996.25%96.14%98.75%0.99490.96140.96710.9711
2Adam320.05015%0.111496.34%0.962696.42%96.10%98.81%0.99660.96100.95140.9600
3Adam960.17565%0.120396.58%0.965996.66%96.53%98.89%0.99550.96530.96530.9700
4Adam320.02515%0.092496.81%0.968196.86%96.77%98.95%0.99720.96770.96580.9712
5Adam400.05020%0.103696.95%0.969296.96%96.88%98.99%0.99650.96880.97010.9741
6Adam560.07530%0.092897.21%0.972597.33%97.17%99.11%0.99670.97170.96580.9713
7Adam960.17565%0.115097.33%0.972997.32%97.26%99.11%0.99440.97260.97630.9789
8Adam400.05020%0.105297.40%0.974397.48%97.38%99.16%0.99550.97380.97690.9796
9Adam400.05020%0.088297.52%0.974397.57%97.30%99.19%0.99690.97300.96690.9727
10Adam560.07530%0.082397.54%0.975297.53%97.51%99.18%0.99680.97510.97650.9797
11Adam960.12545%0.078897.57%0.975697.74%97.39%99.25%0.99770.97390.96760.9733
12Adam560.07530%0.076297.57%0.975897.63%97.53%99.21%0.99780.97530.97410.9781
13Adam320.02515%0.070697.61%0.976597.69%97.61%99.23%0.99840.97610.97040.9757
14Adam560.07530%0.067397.76%0.977497.76%97.72%99.25%0.99840.97720.97500.9790
15Adam560.07530%0.067398.47%0.984598.45%98.45%99.48%0.99720.98450.98620.9878

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 27

ResNet101V2 model correlation results.

MetricBatch sizeDropout ratioModel learn ratio
Loss0.08580.23660.2198
Accuracy0.18300.07980.0957
F10.18720.08040.1002
Precision0.20130.08720.1000
Recall0.17450.07430.1008
Specificity0.20800.09410.1061
AUC−0.2223−0.3566−0.3593
Sensitivity0.17450.07430.1008
IoU Coef0.20100.14050.1986
Dice Coef0.18760.11700.1720
Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union. ResNet152V2: Table 28 reports the ResNet152V2 top-1 results after applying 15 iterations of learning and optimization. The table is sorted in ascending order according to the evaluation accuracy. The best achieved results in all iterations for the loss, accuracy, F1, precision, recall, specificity, AUC, sensitivity, IoU, and dice scores were 0.0935, 96.92%, 0.9694, 97.02%, 96.86%, 99.01%, 0.9972, 0.9686, 0.9679, and 0.9725 respectively while the top-1 results, concerning the accuracy, were 0.1008, 96.92%, 0.9694, 97.02%, 96.86%, 99.01%, 0.9967, 0.9686, 0.9663, and 0.9713 respectively. They were reported by the NAdam parameters optimizer, batch size value of 96, dropout ratio of 0.075, and model learning ratio of 50%. Table 29 reports the correlation between the reported performance metrics and the numeric hyperparameters (i.e. batch size, dropout ratio, and model learn ration).
Table 28

ResNet152V2 learning and optimization top-1 results.

#PO*BS*DR*MLR*LossAccuracyF1PrecisionRecallSpecificityAUC*SensitivityIoU*Dice
1SGD320.40035%0.294488.74%0.886589.23%88.09%96.46%0.98380.88090.88370.9009
2NAdam640.05045%0.278088.88%0.889289.44%88.42%96.52%0.98500.88420.88030.8994
3Adam320.02520%0.205993.12%0.931693.24%93.10%97.75%0.99090.93100.93600.9444
4Adam320.02520%0.169994.15%0.941894.47%93.90%98.17%0.99390.93900.93260.9433
5Adam400.02520%0.194394.22%0.942494.30%94.17%98.10%0.99160.94170.94790.9540
6Adam400.02520%0.131895.46%0.955195.62%95.40%98.54%0.99640.95400.95150.9588
7NAdam960.07550%0.108995.87%0.958796.00%95.73%98.67%0.99710.95730.94930.9585
8Adam320.02520%0.155396.03%0.961196.20%96.02%98.74%0.99320.96020.96560.9696
9NAdam800.07550%0.108796.30%0.962996.53%96.05%98.85%0.99630.96050.95650.9639
10Adam320.02520%0.115896.49%0.965096.66%96.34%98.89%0.99560.96340.95780.9646
11Adam400.02520%0.105396.66%0.965996.70%96.49%98.90%0.99600.96490.96790.9725
12NAdam960.07550%0.093596.84%0.968796.98%96.77%99.00%0.99720.96770.96290.9692
13Adam400.02520%0.097896.84%0.968896.95%96.82%98.99%0.99700.96820.96250.9689
14NAdam720.05035%0.097096.92%0.968896.96%96.80%98.99%0.99690.96800.96490.9706
15NAdam960.07550%0.100896.92%0.969497.02%96.86%99.01%0.99670.96860.96630.9713

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Table 29

ResNet152V2 model correlation results.

MetricBatch sizeDropout ratioModel learn ratio
Loss−0.42000.5295−0.0916
Accuracy0.3011−0.5775−0.0420
F10.2995−0.5862−0.0447
Precision0.3077−0.5748−0.0309
Recall0.2914−0.5960−0.0575
Specificity0.3100−0.5705−0.0272
AUC0.3840−0.57600.0391
Sensitivity0.2914−0.5960−0.0575
IoU Coef0.2161−0.5871−0.1328
Dice Coef0.2380−0.5881−0.1108
Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; AUC, area under curve; IoU, intersection over union.

Top-1 promising results

Table 30 reports Top-1 results of all reported experiments. They are sorted according to the sequence of the reported experiments. The best-achieved loss value was 0.0523, which the DenseNet169 model reported. The best-achieved accuracy value was 98.47% that was reported by the DenseNet169 and ResNet101V2 models. The best achieved F1 value was 0.9849, which the DenseNet169 model reported. The best-achieved precision value was 98.50% that was reported by the DenseNet169 model. The best-achieved recall value was 98.47% that was reported by the DenseNet169 model. The best-achieved specificity value was 99.50% that the DenseNet169 model reported. The best-achieved AUC value was 0.9984, which the ResNet50V2 model reported. The best-achieved sensitivity value was 0.9847, which the DenseNet169 model reported. The best-achieved IoU value was 0.9862, which the ResNet101V2 model reported. The best-achieved dice value was 0.9879, which the DenseNet169 model reported. Figure 9 summarizes the top-1 accuracies.
Table 30

The top-1 results of all reported experiments.

ModelPO*BS*DR*MLR*LossAcc.*F1Prec.*RecallSpec.*AUC*Sen.*IoU*Dice
DenseNet121AdaMax400.07555%0.126897.39%0.973897.44%97.33%99.15%0.99620.97330.97500.9783
DenseNet169SGD880.02590%0.052398.47%0.984998.50%98.47%99.50%0.99830.98470.98600.9879
DenseNet201SGD800.00045%0.115396.69%0.967596.90%96.61%98.97%0.99580.96610.96670.9711
XceptionRMSProp800.52570%0.088897.78%0.978397.90%97.76%99.30%0.99610.97760.97500.9789
MobileNetNAdam560.05025%0.124796.40%0.963696.39%96.33%98.80%0.99520.96330.96640.9706
MobileNetV2RMSProp160.42555%2.514081.54%0.815881.69%81.48%93.91%0.93520.81480.86350.8728
MobileNetV3SmallRMSProp80.075100%27.86080.06%0.799580.50%79.43%93.59%0.92700.79430.81590.8365
MobileNetV3LargeAdaMax160.17520%0.519279.57%0.785183.15%74.50%94.97%0.95310.74500.75500.7931
EfficientNetB0RMSProp640.37580%1.398051.23%0.507052.92%48.72%85.55%0.80030.48720.64780.6724
ResNet50V2AdaGrad160.10030%0.079297.46%0.973797.60%97.14%99.20%0.99840.97140.96350.9703
ResNet101V2Adam560.07530%0.067398.47%0.984598.45%98.45%99.48%0.99720.98450.98620.9878
ResNet152V2NAdam960.07550%0.100896.92%0.969497.02%96.86%99.01%0.99670.96860.96630.9713

Note:

PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; Acc., accuracy; Prec., precision; Spec., specificity; AUC, area under curve; Sen., sensitivity; IoU, intersection over union.

Figure 9

The top-1 accuracies of all reported experiments.

Note: PO, parameters optimizer; BS, batch size; DR, dropout ratio; MLR, model learn ratio; Acc., accuracy; Prec., precision; Spec., specificity; AUC, area under curve; Sen., sensitivity; IoU, intersection over union.

Comparative study

Table 31 shows a comparison between the results of both the proposed OTLD-COVID-19 technique and the other state-of-the-art techniques. The techniques are ordered according to the year of publication. The recorded results show that the proposed OTLD-COVID-19 technique outperforms the four classes’ state-of-the-art compared techniques. Figure 10 depicts a delta comparison between the current study and related works concerning the accuracy. It shows that the current study accuracy exceeds 11 related works accuracies.
Table 31

Comparison the proposed technique with other state-of-the-art techniques.

ResearchClasses #AccuracySensitivitySpecificityPrecisionF1-ScoreAUCRecall
Mukherjee et al. (2020)296.92%94.20%100.0%100.0%97.01%99.22%N/A
Wang, Lin & Wong (2020)392.40%80.00%N/A88.90%N/AN/AN./A
Majeed et al. (2020)4N/A94.52%99.35%N/AN/AN/AN/A
Hammoudi et al. (2020)296.28%97.90%N/A94.80%N/AN/AN/A
Hussain et al. (2021)491.20%91.76%93.48%92.04%90.04%N/A91.90%
Farooq & Hafeez (2020)496.23%100.0%100.0%N/A100.0%N/AN/A
Apostolopoulos & Mpesiana (2020)396.78%98.66%96.46%N/AN/AN/AN/A
Narin, Kaya & Pamuk (2020)298.00%96.00%100.0%100.0%98.00%100.0%N/A
Islam et al. (2020)296.10%N/AN/AN/AN/AN/AN/A
Nayak et al. (2021)289.33%100.0%N/AN/AN/AN/AN/A
Khan, Shah & Bhat (2020)489.60%N/A97.90%93.17%95.61%N/A98.25%
Islam, Islam & Asraf (2020)399.40%99.30%99.20%N/A98.90%99.90%N/A
Aslan et al. (2021)398.70%N/A99.30%N/A98.80%99.00%98.80%
Sethy et al. (2020)395.33%95.33%N/AN/A95.34%N/AN/A
OTLD-COVID-19 Approach498.47%98.47%99.50%98.50%98.49%99.83%98.47%
Figure 10

A delta comparison between the current study and related works.

Conclusion

Early detection of COVID-19 positive cases is essential to prevent the spread of this pandemic and treat affected patients quickly. This study presented an Optimized Transfer Learning-based Approach for Automatic Detection of COVID-19 (OTLD-COVID-19) approach, which adopted the Manta Ray Foraging Optimization (MRFO) algorithm to optimize the parameters and hyperparameters of twelve off-the-shelf CNN architectures. The proposed OTLD-COVID-19 approach aims to aid the radiologists in automating the classification of COVID-19 cases based on chest X-ray images. The OTLD-COVID-19 approach is built upon five essential phases. A four-class real dataset has been constructed from eight public datasets to get a relatively large number of chest X-ray images (=12,933 images). Besides, data augmentation is performed to increase the size of the training set and enhance generalization. The training and testing ratio of the dataset was set as 85% and 15%, respectively. The obtained experimental results showed that the proposed OTLD-COVID-19 approach achieved high-performance metrics that outperformed the compared approaches. To extend this work, Generative Adversarial Networks (GAN) can supplement the lack of training set to improve the performance of the classification process. In expansion, the number of other diseases causing pneumonia may be expanded, and the proposed approach can be utilized to distinguish them from the COVID-19. This study could also be extended to other diseases to help the healthcare system respond more effectively during any possible future pandemic.
  23 in total

1.  Correlation of Chest CT and RT-PCR Testing for Coronavirus Disease 2019 (COVID-19) in China: A Report of 1014 Cases.

Authors:  Tao Ai; Zhenlu Yang; Hongyan Hou; Chenao Zhan; Chong Chen; Wenzhi Lv; Qian Tao; Ziyong Sun; Liming Xia
Journal:  Radiology       Date:  2020-02-26       Impact factor: 11.105

2.  Imaging Profile of the COVID-19 Infection: Radiologic Findings and Literature Review.

Authors:  Ming-Yen Ng; Elaine Y P Lee; Jin Yang; Fangfang Yang; Xia Li; Hongxia Wang; Macy Mei-Sze Lui; Christine Shing-Yen Lo; Barry Leung; Pek-Lan Khong; Christopher Kim-Ming Hui; Kwok-Yung Yuen; Michael D Kuo
Journal:  Radiol Cardiothorac Imaging       Date:  2020-02-13

3.  Ensembles of Deep Learning Models and Transfer Learning for Ear Recognition.

Authors:  Hammam Alshazly; Christoph Linse; Erhardt Barth; Thomas Martinetz
Journal:  Sensors (Basel)       Date:  2019-09-24       Impact factor: 3.576

4.  Clinical Characteristics of Coronavirus Disease 2019 in China.

Authors:  Wei-Jie Guan; Zheng-Yi Ni; Yu Hu; Wen-Hua Liang; Chun-Quan Ou; Jian-Xing He; Lei Liu; Hong Shan; Chun-Liang Lei; David S C Hui; Bin Du; Lan-Juan Li; Guang Zeng; Kwok-Yung Yuen; Ru-Chong Chen; Chun-Li Tang; Tao Wang; Ping-Yan Chen; Jie Xiang; Shi-Yue Li; Jin-Lin Wang; Zi-Jing Liang; Yi-Xiang Peng; Li Wei; Yong Liu; Ya-Hua Hu; Peng Peng; Jian-Ming Wang; Ji-Yang Liu; Zhong Chen; Gang Li; Zhi-Jian Zheng; Shao-Qin Qiu; Jie Luo; Chang-Jiang Ye; Shao-Yong Zhu; Nan-Shan Zhong
Journal:  N Engl J Med       Date:  2020-02-28       Impact factor: 91.245

5.  CNN-based transfer learning-BiLSTM network: A novel approach for COVID-19 infection detection.

Authors:  Muhammet Fatih Aslan; Muhammed Fahri Unlersen; Kadir Sabanci; Akif Durdu
Journal:  Appl Soft Comput       Date:  2020-11-18       Impact factor: 6.725

6.  Application of deep learning techniques for detection of COVID-19 cases using chest X-ray images: A comprehensive study.

Authors:  Soumya Ranjan Nayak; Deepak Ranjan Nayak; Utkarsh Sinha; Vaibhav Arora; Ram Bilas Pachori
Journal:  Biomed Signal Process Control       Date:  2020-11-19       Impact factor: 3.880

7.  Cervical cancer detection in pap smear whole slide images using convNet with transfer learning and progressive resizing.

Authors:  Anant R Bhatt; Amit Ganatra; Ketan Kotecha
Journal:  PeerJ Comput Sci       Date:  2021-02-18

8.  COVID-19 Outbreak Associated with Air Conditioning in Restaurant, Guangzhou, China, 2020.

Authors:  Jianyun Lu; Jieni Gu; Kuibiao Li; Conghui Xu; Wenzhe Su; Zhisheng Lai; Deqian Zhou; Chao Yu; Bin Xu; Zhicong Yang
Journal:  Emerg Infect Dis       Date:  2020-04-02       Impact factor: 6.883

9.  CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest x-ray images.

Authors:  Asif Iqbal Khan; Junaid Latief Shah; Mohammad Mudasir Bhat
Journal:  Comput Methods Programs Biomed       Date:  2020-06-05       Impact factor: 5.428

10.  COVID-Net: a tailored deep convolutional neural network design for detection of COVID-19 cases from chest X-ray images.

Authors:  Linda Wang; Zhong Qiu Lin; Alexander Wong
Journal:  Sci Rep       Date:  2020-11-11       Impact factor: 4.379

View more
  2 in total

1.  A complete framework for accurate recognition and prognosis of COVID-19 patients based on deep transfer learning and feature classification approach.

Authors:  Hossam Magdy Balaha; Eman M El-Gendy; Mahmoud M Saafan
Journal:  Artif Intell Rev       Date:  2022-01-29       Impact factor: 9.588

2.  A generic optimization and learning framework for Parkinson disease via speech and handwritten records.

Authors:  Nada R Yousif; Hossam Magdy Balaha; Amira Y Haikal; Eman M El-Gendy
Journal:  J Ambient Intell Humaniz Comput       Date:  2022-08-26
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.