Javaria Amin1, Muhammad Almas Anjum2, Muhammad Sharif3, Tanzila Saba4, Usman Tariq5. 1. Department of Computer Science, University of Wah, Wah, Pakistan. 2. National University of Technology (NUTECH), IJP Road Islamabad, Pakistan. 3. Department of Computer Science, COMSATS University Islamabad - Wah Campus, Wah Cantt, Pakistan, 4740, Pakistan. 4. Artificial Intelligence and Data Analytics (AIDA) Lab CCIS Prince Sultan University, Riyadh, Saudi Arabia. 5. College of Computer Engineering and Sciences, Prince Sattam Bin Abdulaziz University, Alkharj, Saudi Arabia.
Abstract
Coronavirus19 is caused due to infection in the respiratory system. It is the type of RNA virus that might infect animal and human species. In the severe stage, it causes pneumonia in human beings. In this research, hand-crafted and deep microscopic features are used to classify lung infection. The proposed work consists of two phases; in phase I, infected lung region is segmented using proposed U-Net deep learning model. The hand-crafted features are extracted such as histogram orientation gradient (HOG), noise to the harmonic ratio (NHr), and segmentation based fractal texture analysis (SFTA) from the segmented image, and optimum features are selected from each feature vector using entropy. In phase II, local binary patterns (LBPs), speeded up robust feature (Surf), and deep learning features are extracted using a pretrained network such as inceptionv3, ResNet101 from the input CT images, and select optimum features based on entropy. Finally, the optimum selected features using entropy are fused in two ways, (i) The hand-crafted features (HOG, NHr, SFTA, LBP, SURF) are horizontally concatenated/fused (ii) The hand-crafted features (HOG, NHr, SFTA, LBP, SURF) are combined/fused with deep features. The fused optimum features vector is passed to the ensemble models (Boosted tree, bagged tree, and RUSBoosted tree) in two ways for the COVID19 classification, (i) classification using fused hand-crafted features (ii) classification using fusion of hand-crafted features and deep features. The proposed methodology is tested /evaluated on three benchmark datasets. Two datasets employed for experiments and results show that hand-crafted & deep microscopic feature's fusion provide better results compared to only hand-crafted fused features.
Coronavirus19 is caused due to infection in the respiratory system. It is the type of RNA virus that might infect animal and human species. In the severe stage, it causes pneumonia in human beings. In this research, hand-crafted and deep microscopic features are used to classify lung infection. The proposed work consists of two phases; in phase I, infected lung region is segmented using proposed U-Net deep learning model. The hand-crafted features are extracted such as histogram orientation gradient (HOG), noise to the harmonic ratio (NHr), and segmentation based fractal texture analysis (SFTA) from the segmented image, and optimum features are selected from each feature vector using entropy. In phase II, local binary patterns (LBPs), speeded up robust feature (Surf), and deep learning features are extracted using a pretrained network such as inceptionv3, ResNet101 from the input CT images, and select optimum features based on entropy. Finally, the optimum selected features using entropy are fused in two ways, (i) The hand-crafted features (HOG, NHr, SFTA, LBP, SURF) are horizontally concatenated/fused (ii) The hand-crafted features (HOG, NHr, SFTA, LBP, SURF) are combined/fused with deep features. The fused optimum features vector is passed to the ensemble models (Boosted tree, bagged tree, and RUSBoosted tree) in two ways for the COVID19 classification, (i) classification using fused hand-crafted features (ii) classification using fusion of hand-crafted features and deep features. The proposed methodology is tested /evaluated on three benchmark datasets. Two datasets employed for experiments and results show that hand-crafted & deep microscopic feature's fusion provide better results compared to only hand-crafted fused features.
The World Health organization has named Coronavirus19 (COVID‐19) a global pandemic disease (WHO).The severity of this disease may disrupt the respiratory system (Holshue et al., 2020). COVID‐19 is also recognized as a deadly disease having a fatality rate of 2%. Massive damage of alveolar and the respiratory system's failure are the main causes of death in severe circumstances of this disease (Saba, Abunadi, Shahzad, & Khan, 2021). RT‐ PCR is the laboratory technique for the diagnosis of this disorder. For early detection of this disease, RT‐PCR is not effective because its testing process is time‐consuming which may cause a delay in the proper treatment of the patients (Fang et al., 2020; Khan et al., 2021). The effective imaging protocol is Computed Tomography (CT) which is noninvasive (Huang et al., 2020). This is an essential diagnostic tool for early screening; however, there is difficulty in differentiating the COVID‐19 features from other pneumonia types. It is beneficial to diagnose and quarantine the patient and to control the spreading of the disease. The task of object detection and classification of images is the main concern in deep learning. Artificial intelligence is being applied on data to recognize patterns and their analysis (Rehman, Saba, et al., 2021; Rehman, Sabad, et al., 2021; Wang, Hu, et al., 2020). Therefore, it can be more beneficial for COVID‐19 detection and classification. Different AI methods are utilized for COVID‐19 detection. However, there is still a gap in this domain because small lesion regions are detected as a normal region and existing algorithms failed when lesion appears at the lungs region's border (Haimed, Saba, Albasha, Rehman, & Kolivand, 2021; Saba, 2021). Optimal features extraction and selection is another challenge. At the same time noisy features also reduce the overall model accuracy(Acharya et al., 2019; Albahri et al., 2020a, 2020b). Therefore, a new approach focused on semantic segmentation and features analysis based on the hand‐crafted and deep fusion features is discussed in this study. The core contributed steps of the proposed method include:The affected lung region is segmented using a modified U‐Net deep learning model.Extracted hand‐crafted deep features and selected optimized features using entropy are fused serially and supplied to the ensemble learning.
RELATED WORK
Artificial intelligence (AI) approaches are commonly utilized in biomedical research (Kang et al., 2020; Saba, 2019; Saba, Bokhari, Sharif, Yasmin, & Raza, 2018; Salman, Ahmed, Khan, Raza, & Latif, 2017; Wang, Muhammad, et al., 2020; Wang, Sun, et al., 2020; Wang, Tang, et al., 2019; Wang, Zhang, et al., 2019). AI methods are used in different applications that is, object detection, segmentation, and classification (Deepa, Devi,, & Technology, 2011). COVID19 patients are suffered from pneumonia due to the RNA virus that infects the lung region(Lai, Shih, Ko, Tang, & Hsueh, 2020). Deep learning is widely used to detect lung‐infected regions (Shan et al., 2020). Deep learning models are being used in three different ways such as pretrained model as fine‐tuning, without fine‐tuning, and trained from scratch. In the pretrained models, inceptionv3, VGG16, squeeze Net, Mobilenetv2, and ResNet‐50 are widely used for pneumonia detection (Kassani, Kassasni, Wesolowski, Schneider, & Deters, 2020). In the literature, a massive amount of work is done for COVID‐19 detection; however, still, there is a need for an optimized features extraction/selection approach which helps for accurate detection of the COVID‐19 (Albahri et al., 2020a, 2020b). Therefore, in this research hand‐crafted and deep features are extracted and fused for better classification of COVID‐19. The existing literature on COVID‐19 are summarized in Table 1.
TABLE 1
Overview of the most recent COVID‐19 work
References
Methods
Data
Results
Shan et al. (2020)
VB neural network
300 patients of COVID‐19
91.6% ± 10.0% dice coefficient
Kassani et al. (2020)
Transfer learning models
117 chest X‐ray images
98% accuracy
Horry et al. (2020)
Pretrained VGG model
300 patients of COVID‐19
86% precision rate
Overview of the most recent COVID‐19 work
PROPOSED MODEL FOR DETECTION AND CLASSIFICATION OF CORONAVIRUS19
In this retrospective study, U‐Net based semantic deep learning model is proposed for lung region segmentation, which is infected by an RNA virus. Later, extracted hand‐crafted and deep features from the original CT images and segmented images. The redundant features affect the overall model accuracy; therefore, discriminant features are selected using entropy. The selected features are serially concatenated/fused for accurate classification of the COVID19. The proposed model architecture is presented in Figure 1.
FIGURE 1
Proposed model for segmentation and classification of healthy/COVID‐19 images
Proposed model for segmentation and classification of healthy/COVID‐19 images
Segmentation of the infected lung region using U‐Net deep convolutional neural network
The CT images of size 224 × 224 × 1 are passed to the U‐Net (He, Zhang, Ren, & Sun, 2015) semantic segmentation deep learning model. The proposed model contains 46 layers in which 01 input, 15 convolutional (encoder), 17 ReLU, 03 max‐pooling, 02 drop‐out, 04, 01 softmax, and classification layers. The learning patterns of the last encoder and decoder section layers are visually presented in Figure 2.
FIGURE 2
U‐Net model for COVID‐19 segmentation
U‐Net model for COVID‐19 segmentationThe bridge contains two sets of convolutional and ReLU layers. The encoder‐depth specifies the depth of encoder and decoder sub‐networks. Thus in this work, encoder depth is set as 3. The encoder section contains four stages, in which two sets of the convolutional and ReLU layers. Each stage contains one Maxpooling layer to perform down‐sampling and a 50% dropout layer is also used in the fourth stage to prevent the over‐fitting problem. The decoder section also contains four stages, in which each stage contains two sets of the convolutional, ReLU and transpose convolutional layer to perform up‐sampling. The final convolutional layer was followed by the softmax and segmentation layers which performed semantic segmentation. The model is trained from scratch and the selected configuration parameters of the U‐Net model are mentioned in Table 2. At the same time, the segmented images with ground truth are shown in Figure 3.
TABLE 2
U‐Net configuration parameters
Input size
Encoder depth
MaxEpoch
Learning rate
224 × 224 × 1
3
40
0.001
FIGURE 3
Segmentation results with ground truth (a) input images (b) segmentation by proposed method (c) manually ground truth images
U‐Net configuration parametersSegmentation results with ground truth (a) input images (b) segmentation by proposed method (c) manually ground truth images
Features extraction
In the proposed method, hand‐crafted and deep features are extracted for COVID‐19 classification.
Hand crafted features extraction
The five hand‐crafted features such as noise to the harmonic ratio (NHr), histogram orientation gradient (HOG), local binary patterns (LBPs), SFTA, and speeded up robust features (SURF) are extracted. The best features are selected based on the maximum scores using entropy and fused serially.
Noise to harmonic ratio features
The NHr (Kim, Moreau, & Sikora, 2006) features is extracted from the segmented region. NHr is a strong feature to identify the lesion region's variation because lung lesions have variable shape and size. The NHr features are represented mathematically as follows:
where N (number of sample), M represents the maximum lag. The length of NHr features vector is 1 × 128 and graphically illustrated in Figure 4.
FIGURE 4
Noise to the harmonic ratio (NHr) features (a) input image (b) NHr
Noise to the harmonic ratio (NHr) features (a) input image (b) NHr
Speeded up robust features
Extracted speeded up robust features (SURF) features in terms of points, orientation, and scale from the input CT images having 1 × 96 length that is used for the detection of interest points as illustrated in Figure 5.
FIGURE 5
SURF features (a) original image (b) SURF features
SURF features (a) original image (b) SURF features
Histogram orientation gradient
The lesion regions have an irregular shape. Therefore, HOG (shape‐based) features are extracted from the segmented lesion region having a vector length is 1 × 8,100 and are shown in Figure 6.
FIGURE 6
Histogram orientation gradient (HOG) features (a) original image (b) HOG
Histogram orientation gradient (HOG) features (a) original image (b) HOG
Local binary pattern
The lung CT images having variant texture; therefore, LBP features are extracted from gray‐level images. The feature vector length is 1 × 59 and graphically shown in Figure 7.
FIGURE 7
Local binary patterns (LBP) of the original image
Local binary patterns (LBP) of the original image
SFTA
SFTA features are extracted from the segmented lesion region, which creates its own segment pattern and helps classify. The length of the feature vector is 1 × 22 and graphically shown in Figure 8.
FIGURE 8
SFTA features
SFTA features
DCNN features
Extracted deep features using pretrained deep learning models that is, ResNet 101 and inceptionv3. The ResNet101 contains 347 layers in which 01 input, 107 convolutional, 99 ReLU, 103 batch‐normalization, 01 maxpooling, 32 addition layers, 01 pool average global layer, fully connected, 01 softmax, & classification layer. While inceptionv3 contains 315 layers in which input, 96 convolutional, 94 batch‐normalization, 94 ReLU, 04 maxpooling, 09 average pooling layers, 14 depth‐concatenations, fully‐connected, classification& softmax layer. The fully connected layer named as FC1000 of Resnet101 and the prediction layer of inceptionv3 with 1 × 1000 features vector length is selected for classification and graphically plot in Figure 9.
FIGURE 9
DCNN features
DCNN features
Features selection and fusion
The features selection approach is presented using entropy as mentioned in Equation (2). The hand‐crafted and deep features are extracted from input images. The complete process of the feature's extraction, selection, and fusion is shown in Figure 10.
FIGURE 10
Features extraction, selection, and fusion process for COVID‐19 classification
Features extraction, selection, and fusion process for COVID‐19 classificationThe length of extracted features is such as fNHr1 × 128, fHOG1 × 8,100, fSFTA1 × 22, fSURF1 × 96, fResNet1011 × 1,000, and fInceptionv31 × 1,000. The best features are selected based on the maximum score using entropy.H
denotes entropy, N represents the number of features, and n shows features at the same score level.The features are selected based on maximum score such as, 15 NHr features out of 128, 3,935 HOG features out of 8,100, 15 SFTA features out of 22, 90 SURF features out of 96, and 100 deep features out of 1,000 (ResNet101 and Inceptionv3). The selected score‐based features vectors are denoted as 1 × a, 1 × b, 1 × c, 1 × d, 1 × e, 1 × f. The selected features are concatenated/fused horizontally and denoted as FV1 × z .Where,
andThe fused features vector length is 1 × 4256 in which the last column represents the class labels. The final features vectors with class labels are passed to the ensemble classifiers.
Classification
The vector of resultant fused feature is fed to classifiers. The ensemble classifiers are used with three kernels such as a boosted tree (Freund, 2009), bagged tree (Breiman, 1996), and RUSBoosted tree (Seiffert, Khoshgoftaar, Van Hulse, & Napolitano, 2008). The ensemble classifiers are built by combining the week decision tree classifiers to create a powerful prediction model. Brief overview of the selected classifiers with the learning parameters is given in Table 3.
TABLE 3
Brief overview of ensemble classifiers
Model type
Ensemble approach
Learner
Maximum splits
Total learners
Rate of the learning
Training time
Boosted Tree
AdaBoost
Decision tree
20
30
0.1
380.3 s
Bagged Tree
Bag
877
64.689 s
RUSBoosted Tree
RUSBoost
20
189.98 s
Brief overview of ensemble classifiers
BENCHMARK DATASETS
The proposed research is evaluated using three benchmark datasets. Dataset I, COVID‐19 segmentation data consists of the 40 patients, in which 100 slices of CT images with ground truth in NII format. Besides, this dataset out of 829 slices includes 373 positive cases (COVID‐19). Dataset II, consists of the Chinese Hospital CT imaging data of 216 patients. The patient's data contains 316 positive and 700 negative slices (Zhao, Zhang, He, & Xie, 2020). Dataset III includes 30 patients of POF Hospital, in which 2000 positive slices are included out of 4,500. The classified radiologists also created the ground truth. The description of training and testing images is given in Table 4.
TABLE 4
Training and testing data on benchmark datasets
Dataset
Total images
Training and testing images
Classification
Dataset I
829
456 negative images and 373 positive images
10 fold cross validation (means divide whole data into 10 sets, one randomly selected set is utilized for testing and remaining nine for training. Same methodology is repeated in all 10 iterations.
Dataset II
1,016
316 positive and 700 negative slices
Training and testing data on benchmark datasets
Experimentation
The proposed model is implemented on MATLAB core i7 CPU with Nvidia Graphic card. Two experiments are performed to test the proposed method performance. In Experiment 1, pixel‐based segmentation results are evaluated with ground truth. In Experiment 2, classifications are performed using hand‐crafted fused features and hand‐crafted and deep features fusion. In this experiment, training/testing data is divided on cross‐validation of 10‐fold. The performance of the proposed approach is assessed on various types of performance measurements described in Table 5.
TABLE 5
Performance measurements
Measures
Formulas
IoU
ϑϑ+ρ+δ here ϑ denotes true positive, false positive (ρ), and false negative (δ)
Accuracy
ϑ+∃ϑ+ρ+δ+∃ ∃ represents the true negative
Sensitivity
ϑϑ+δ
Specificity
∃ρ+∃
Performance measurements
Experiment #1 (segmentation of lung region affected by COVID19)
In this experiment, segmented lesion region is evaluated with ground truth. IoU is mathematically expressed as:
The segmentation model is trained on selected parameters, which are finalized after the extensive experiments alluded in Table 6.
TABLE 6
Experiment for the selection of optimal parameters
Encoder depth
Optimizer
Epochs
Learning rate
Training accuracy
2
Adam
30
0.002
0.96
3
Sgdm
40
0.001
0.98
3
RmsProp
100
0.001
0.95
Note: To prominant the selected parameters used and the results acheived from the proposed method are bold.
Experiment for the selection of optimal parametersNote: To prominant the selected parameters used and the results acheived from the proposed method are bold.Table 6, shows learning parameters of the proposed segmentation model where different values are selected for training of model and to analayze the accuracy. The different kinds of optimizers such as sgdm, adam, and rmsProp are employed with 40, 30, and 100 epochs, respectively. We observed that learning rate is 0.001, 40 epochs, and encoder depth 3 with sgdm optimizer provide greater training accuracy than other optimizers. These parameters are reflected in italic highlighted font and utilized for further experiments. The quantitative segmentation results are reflected in the Table 7.
TABLE 7
Segmentation outcomes on benchmark datasets
Dataset
Global ACC
Mean ACC
Mean IoU
Weighted IoU
BF mean score
COVID‐19 segmentation
0.98
0.88
0.87
0.96
0.89
POF hospital
1.00
1.00
1.00
1.00
1.00
Segmentation outcomes on benchmark datasetsTable 3 shows that the modified U‐Net segmentation model achieves 0.96 IoU on the COVID‐19 segmentation dataset and 1.00 IoU on the POF Hospital dataset. The segmentation results of the lung affected region with ground truth are shown in Figures 11 and 12.
FIGURE 11
Segmentation outcome on COVID‐19 segmentation dataset
FIGURE 12
Segmentation outcome on POF Hospital dataset
Segmentation outcome on COVID‐19 segmentation datasetSegmentation outcome on POF Hospital dataset
Experiment #2 (classification based on hand‐crafted fused features)
The classification experiment is performed by extraction of the different hand‐crafted features such as NHr, SFTA (Costa, Humpire‐Mamani, & Traina, 2012), HOG(Dalal & Triggs, 2005), LBP (Ojala, Pietikainen, & Maenpaa, 2002), and SURF (Bay, Ess, Tuytelaars, & Van Gool, 2008), and at the same time informative features are selected using entropy and fused serially (1 × 4055) vector length. That is provided to the ensemble method with three powerful learning models such as boosted, bagged, and the Rusboosted tree. The ensemble model prediction is graphically presented in the Figure 13.
FIGURE 13
Ensemble prediction model based on hand crafted features vector
Ensemble prediction model based on hand crafted features vectorThe quantitative classification outcomes are presented in Table 8, 9, 10.
TABLE 8
COVID‐19 classification on COVID‐19 segmentation dataset
Classification model
Sensitivity
Specificity
Accuracy
Boosted tree
0.90
0.91
86.3%
Bagged tree
0.98
0.97
96.8%
RUSBoosted tree
1.0
0.97
98.17%
TABLE 9
COVID‐19 classification on Chinese hospital dataset
Classification model
Sensitivity
Specificity
Accuracy
Boosted Tree
0.97
0.92
92.03%
Bagged Tree
0.90
0.70
89.52%
RUSBoosted Tree
0.99
1.0
98.97%
TABLE 10
COVID‐19 classification on POF hospital dataset
Classification model
Sensitivity
Specificity
Accuracy
Boosted Tree
1.0
1.0
99.77%
Bagged Tree
0.96
0.99
96.06%
RUSBoosted Tree
0.99
0.93
93.85%
COVID‐19 classification on COVID‐19 segmentation datasetCOVID‐19 classification on Chinese hospital datasetCOVID‐19 classification on POF hospital datasetThe results in Tables 8, 9, 10 shows that RUSBoosted tree ensemble model provided better results in term of ACC, such as 98.17% on COVID‐19 segmentation dataset and 98.97% on Chinese Hospital dataset. At the same time, Boosted Tree ensemble model provides the highest accuracy on the POF hospital dataset.
Experiment 2.1 horizontally concatenation of the hand crafted and deep convolutional neural network
Extracted deep features are using Deep Convolutional Neural Network (DCNN) such as ResNET101(He, Zhang, Ren, & Sun, 2016) and Inception v3(Szegedy, Vanhoucke, Ioffe, Shlens, & Wojna, 2016) where length of the deep extracted features vector is 1 × 1,000. One hundred active features are selected from each feature vector using entropy and selected deep features. They are serially fused having a vector length 1 × 200 and further fused with hand‐crafted features vector length of 1 × 4,055. The length of final fused features is 1 × 4,255 which is passed to the ensemble learning classifiers for classification. The results with class labels are presented in Figure 14.
FIGURE 14
Confusion matrix using ensemble methods (a) bagged tree, (b) RUSBoosted tree, (c) Boosted tree
Confusion matrix using ensemble methods (a) bagged tree, (b) RUSBoosted tree, (c) Boosted treeIn Figure 14, d denotes the disease class and h represents the healthy. The model prediction is performed using the combination of classical and deep features as illustrated in Figure 15.
FIGURE 15
An ensemble prediction model based on hand‐crafted and deep features fusion
An ensemble prediction model based on hand‐crafted and deep features fusionThe empirical results are provided in Tables 11, 12, 13.
TABLE 11
COVID‐19 classification on COVID‐19 segmentation dataset
Classification model
Sensitivity
Specificity
Accuracy
Boosted Tree
0.99
1.0
99.09%
Bagged Tree
1.00
1.00
100%
RUSBoosted Tree
0.98
0.97
96.8%
TABLE 12
COVID‐19 classification on Chinese hospital dataset
Classification model
Sensitivity
Specificity
Accuracy
Boosted Tree
0.98
1.0
98.52%
Bagged Tree
0.97
0.92
92.14%
RUSBoosted Tree
0.89
0.97
88.61%
TABLE 13
COVID‐19 classification on POF hospital dataset
Classification model
Sensitivity
Specificity
Accuracy
Boosted Tree
0.97
1.0
97.72%
Bagged Tree
0.91
0.96
89.73%
RUSBoosted Tree
0.96
0.92
91.25%
COVID‐19 classification on COVID‐19 segmentation datasetCOVID‐19 classification on Chinese hospital datasetCOVID‐19 classification on POF hospital datasetThe results in Tables 11, 12, 13 show that the ensemble boosted tree, achieved 99.09% ACC on the COVID‐19 segmentation dataset, 98.52% ACC on the Chinese Hospital dataset, and 97.72% ACC on the POF Hospital dataset. While ensemble bagged tree obtained 100% ACC on segmentation of COVID‐19 dataset. The ratio among the AUC and ROC is also plotted in Figure 16.
FIGURE 16
AUC across ROC of ensemble methods (a) Bagged tree, (b) Boosted tree, (c) RUSBoosted tree
AUC across ROC of ensemble methods (a) Bagged tree, (b) Boosted tree, (c) RUSBoosted treeIn Figure 16, the maximum achieved AUC is 1.00 using the bagged tree ensemble method, 0.95 AUC on the Boosted tree ensemble method, where 0.97 AUC on the RUSBoosted tree ensemble. The classification results are analyzed by ensembling different features such as texture, shape, NHr, and deep features. The findings of the proposed procedure are contrasted with the recent studies published alluded to in Table 14.
TABLE 14
Comparison of the proposed method with existing works
References
Datasets
Method
Results accuracy%
Yang et al. (2020)
UCSD‐AI4H/COVID‐CT
ResNet‐50
0.89
Burgos‐Artizzu (2020)
Inception
83.5 ± 2.0
Z. Wang, Liu, and Dou (2020)
COVID‐Net
78.69 ± 1.54
Ewen and Khan (2020)
DenseNet169
0.8621 ± 0.0474
Proposed method
Fusion of hand crafted and deep features
0.9852
Comparison of the proposed method with existing worksThe proposed framework is contrasted with four latest frameworks that is, (Burgos‐Artizzu, 2020; Ewen & Khan, 2020; Wang, Liu, et al., 2020; Yang et al., 2020). The existing published methods classified the COVID‐19 CT images using different pretrained CNN models, these methods achieved a maximum of 0.89 accuracy. However, this research provides a unique framework based on the deeper analysis of the deep and hand crafted features with more prominent features selection using entropy. The proposed approach provides a significant improvement of 0.9852 accuracy for the classification of COVID‐10 lung CT images. This approach might be utilized in medical domain to analyze the more informative features for disease classification.In this study, after comprehensive features analysis, it is observed that serially fusion of the hand‐crafted and deep features are provided a good effect on the model accuracy.
CONCLUSION
COVID‐19 segmentation and classification are an intricate task at an early stage. The accurate COVID‐19 segmentation is a crucial task because lesion regions appear at the boundaries that are difficult to segment using classical segmentation methods. Therefore, a deep learning model can be used to accurately segmentation the lesion region that appears at the borders. In this research, the modified U‐Net model is employed for segmentation. The proposed segmentation model achieved 0.98 global ACC, 0.96 weighted IoU on the COVID‐19 segmentation dataset, and 0.98 global ACC, 1.00 IoU on the POF Hospital dataset. The experimental discussion concludes that the fusion of the hand‐crafted and deep features provided competent results as compared to the fusion of hand‐crafted features only. The models achieved ACC of 99.09% on the Boosted Tree ensemble, 100% on the Bagged tree ensemble, and 96.8% RUSBoosted tree ensemble model.
CONFLICT OF INTEREST
All authors declared that there is no conflict of interest.
AUTHOR CONTRIBUTIONS
All authors are contributed equally in this manuscript.
Authors: O S Albahri; A A Zaidan; A S Albahri; B B Zaidan; Karrar Hameed Abdulkareem; Z T Al-Qaysi; A H Alamoodi; A M Aleesa; M A Chyad; R M Alesa; L C Kem; Muhammad Modi Lakulu; A B Ibrahim; Nazre Abdul Rashid Journal: J Infect Public Health Date: 2020-07-01 Impact factor: 3.718
Authors: Michelle L Holshue; Chas DeBolt; Scott Lindquist; Kathy H Lofy; John Wiesman; Hollianne Bruce; Christopher Spitters; Keith Ericson; Sara Wilkerson; Ahmet Tural; George Diaz; Amanda Cohn; LeAnne Fox; Anita Patel; Susan I Gerber; Lindsay Kim; Suxiang Tong; Xiaoyan Lu; Steve Lindstrom; Mark A Pallansch; William C Weldon; Holly M Biggs; Timothy M Uyeki; Satish K Pillai Journal: N Engl J Med Date: 2020-01-31 Impact factor: 91.245
Authors: Sara Hosseinzadeh Kassania; Peyman Hosseinzadeh Kassanib; Michal J Wesolowskic; Kevin A Schneidera; Ralph Detersa Journal: Biocybern Biomed Eng Date: 2021-06-05 Impact factor: 4.314
Authors: Javaria Amin; Muhammad Almas Anjum; Muhammad Sharif; Seifedine Kadry; Ahmed Nadeem; Sheikh F Ahmad Journal: Diagnostics (Basel) Date: 2022-03-27
Authors: Tahira Nazir; Marriam Nawaz; Ali Javed; Khalid Mahmood Malik; Abdul Khader Jilani Saudagar; Muhammad Badruddin Khan; Mozaherul Hoque Abul Hasanat; Abdullah AlTameem; Mohammad AlKathami Journal: Microsc Res Tech Date: 2022-02-23 Impact factor: 2.893