Literature DB >> 33964096

An intelligence design for detection and classification of COVID19 using fusion of classical and convolutional neural network and improved microscopic features selection approach.

Javaria Amin1, Muhammad Almas Anjum2, Muhammad Sharif3, Tanzila Saba4, Usman Tariq5.   

Abstract

Coronavirus19 is caused due to infection in the respiratory system. It is the type of RNA virus that might infect animal and human species. In the severe stage, it causes pneumonia in human beings. In this research, hand-crafted and deep microscopic features are used to classify lung infection. The proposed work consists of two phases; in phase I, infected lung region is segmented using proposed U-Net deep learning model. The hand-crafted features are extracted such as histogram orientation gradient (HOG), noise to the harmonic ratio (NHr), and segmentation based fractal texture analysis (SFTA) from the segmented image, and optimum features are selected from each feature vector using entropy. In phase II, local binary patterns (LBPs), speeded up robust feature (Surf), and deep learning features are extracted using a pretrained network such as inceptionv3, ResNet101 from the input CT images, and select optimum features based on entropy. Finally, the optimum selected features using entropy are fused in two ways, (i) The hand-crafted features (HOG, NHr, SFTA, LBP, SURF) are horizontally concatenated/fused (ii) The hand-crafted features (HOG, NHr, SFTA, LBP, SURF) are combined/fused with deep features. The fused optimum features vector is passed to the ensemble models (Boosted tree, bagged tree, and RUSBoosted tree) in two ways for the COVID19 classification, (i) classification using fused hand-crafted features (ii) classification using fusion of hand-crafted features and deep features. The proposed methodology is tested /evaluated on three benchmark datasets. Two datasets employed for experiments and results show that hand-crafted & deep microscopic feature's fusion provide better results compared to only hand-crafted fused features.
© 2021 Wiley Periodicals LLC.

Entities:  

Keywords:  U-Net; ensemble methods; entropy; fusion; hand crafted features; healthcare; public health

Mesh:

Year:  2021        PMID: 33964096      PMCID: PMC8237066          DOI: 10.1002/jemt.23779

Source DB:  PubMed          Journal:  Microsc Res Tech        ISSN: 1059-910X            Impact factor:   2.893


INTRODUCTION

The World Health organization has named Coronavirus19 (COVID‐19) a global pandemic disease (WHO).The severity of this disease may disrupt the respiratory system (Holshue et al., 2020). COVID‐19 is also recognized as a deadly disease having a fatality rate of 2%. Massive damage of alveolar and the respiratory system's failure are the main causes of death in severe circumstances of this disease (Saba, Abunadi, Shahzad, & Khan, 2021). RT‐ PCR is the laboratory technique for the diagnosis of this disorder. For early detection of this disease, RT‐PCR is not effective because its testing process is time‐consuming which may cause a delay in the proper treatment of the patients (Fang et al., 2020; Khan et al., 2021). The effective imaging protocol is Computed Tomography (CT) which is noninvasive (Huang et al., 2020). This is an essential diagnostic tool for early screening; however, there is difficulty in differentiating the COVID‐19 features from other pneumonia types. It is beneficial to diagnose and quarantine the patient and to control the spreading of the disease. The task of object detection and classification of images is the main concern in deep learning. Artificial intelligence is being applied on data to recognize patterns and their analysis (Rehman, Saba, et al., 2021; Rehman, Sabad, et al., 2021; Wang, Hu, et al., 2020). Therefore, it can be more beneficial for COVID‐19 detection and classification. Different AI methods are utilized for COVID‐19 detection. However, there is still a gap in this domain because small lesion regions are detected as a normal region and existing algorithms failed when lesion appears at the lungs region's border (Haimed, Saba, Albasha, Rehman, & Kolivand, 2021; Saba, 2021). Optimal features extraction and selection is another challenge. At the same time noisy features also reduce the overall model accuracy(Acharya et al., 2019; Albahri et al., 2020a, 2020b). Therefore, a new approach focused on semantic segmentation and features analysis based on the hand‐crafted and deep fusion features is discussed in this study. The core contributed steps of the proposed method include: The affected lung region is segmented using a modified U‐Net deep learning model. Extracted hand‐crafted deep features and selected optimized features using entropy are fused serially and supplied to the ensemble learning.

RELATED WORK

Artificial intelligence (AI) approaches are commonly utilized in biomedical research (Kang et al., 2020; Saba, 2019; Saba, Bokhari, Sharif, Yasmin, & Raza, 2018; Salman, Ahmed, Khan, Raza, & Latif, 2017; Wang, Muhammad, et al., 2020; Wang, Sun, et al., 2020; Wang, Tang, et al., 2019; Wang, Zhang, et al., 2019). AI methods are used in different applications that is, object detection, segmentation, and classification (Deepa, Devi,, & Technology, 2011). COVID19 patients are suffered from pneumonia due to the RNA virus that infects the lung region(Lai, Shih, Ko, Tang, & Hsueh, 2020). Deep learning is widely used to detect lung‐infected regions (Shan et al., 2020). Deep learning models are being used in three different ways such as pretrained model as fine‐tuning, without fine‐tuning, and trained from scratch. In the pretrained models, inceptionv3, VGG16, squeeze Net, Mobilenetv2, and ResNet‐50 are widely used for pneumonia detection (Kassani, Kassasni, Wesolowski, Schneider, & Deters, 2020). In the literature, a massive amount of work is done for COVID‐19 detection; however, still, there is a need for an optimized features extraction/selection approach which helps for accurate detection of the COVID‐19 (Albahri et al., 2020a, 2020b). Therefore, in this research hand‐crafted and deep features are extracted and fused for better classification of COVID‐19. The existing literature on COVID‐19 are summarized in Table 1.
TABLE 1

Overview of the most recent COVID‐19 work

ReferencesMethodsDataResults
Shan et al. (2020)VB neural network300 patients of COVID‐1991.6% ± 10.0% dice coefficient
Kassani et al. (2020)Transfer learning models117 chest X‐ray images98% accuracy
Horry et al. (2020)Pretrained VGG model300 patients of COVID‐1986% precision rate
Overview of the most recent COVID‐19 work

PROPOSED MODEL FOR DETECTION AND CLASSIFICATION OF CORONAVIRUS19

In this retrospective study, U‐Net based semantic deep learning model is proposed for lung region segmentation, which is infected by an RNA virus. Later, extracted hand‐crafted and deep features from the original CT images and segmented images. The redundant features affect the overall model accuracy; therefore, discriminant features are selected using entropy. The selected features are serially concatenated/fused for accurate classification of the COVID19. The proposed model architecture is presented in Figure 1.
FIGURE 1

Proposed model for segmentation and classification of healthy/COVID‐19 images

Proposed model for segmentation and classification of healthy/COVID‐19 images

Segmentation of the infected lung region using U‐Net deep convolutional neural network

The CT images of size 224 × 224 × 1 are passed to the U‐Net (He, Zhang, Ren, & Sun, 2015) semantic segmentation deep learning model. The proposed model contains 46 layers in which 01 input, 15 convolutional (encoder), 17 ReLU, 03 max‐pooling, 02 drop‐out, 04, 01 softmax, and classification layers. The learning patterns of the last encoder and decoder section layers are visually presented in Figure 2.
FIGURE 2

U‐Net model for COVID‐19 segmentation

U‐Net model for COVID‐19 segmentation The bridge contains two sets of convolutional and ReLU layers. The encoder‐depth specifies the depth of encoder and decoder sub‐networks. Thus in this work, encoder depth is set as 3. The encoder section contains four stages, in which two sets of the convolutional and ReLU layers. Each stage contains one Maxpooling layer to perform down‐sampling and a 50% dropout layer is also used in the fourth stage to prevent the over‐fitting problem. The decoder section also contains four stages, in which each stage contains two sets of the convolutional, ReLU and transpose convolutional layer to perform up‐sampling. The final convolutional layer was followed by the softmax and segmentation layers which performed semantic segmentation. The model is trained from scratch and the selected configuration parameters of the U‐Net model are mentioned in Table 2. At the same time, the segmented images with ground truth are shown in Figure 3.
TABLE 2

U‐Net configuration parameters

Input sizeEncoder depthMaxEpochLearning rate
224 × 224 × 13400.001
FIGURE 3

Segmentation results with ground truth (a) input images (b) segmentation by proposed method (c) manually ground truth images

U‐Net configuration parameters Segmentation results with ground truth (a) input images (b) segmentation by proposed method (c) manually ground truth images

Features extraction

In the proposed method, hand‐crafted and deep features are extracted for COVID‐19 classification.

Hand crafted features extraction

The five hand‐crafted features such as noise to the harmonic ratio (NHr), histogram orientation gradient (HOG), local binary patterns (LBPs), SFTA, and speeded up robust features (SURF) are extracted. The best features are selected based on the maximum scores using entropy and fused serially.

Noise to harmonic ratio features

The NHr (Kim, Moreau, & Sikora, 2006) features is extracted from the segmented region. NHr is a strong feature to identify the lesion region's variation because lung lesions have variable shape and size. The NHr features are represented mathematically as follows: where N (number of sample), M represents the maximum lag. The length of NHr features vector is 1 × 128 and graphically illustrated in Figure 4.
FIGURE 4

Noise to the harmonic ratio (NHr) features (a) input image (b) NHr

Noise to the harmonic ratio (NHr) features (a) input image (b) NHr

Speeded up robust features

Extracted speeded up robust features (SURF) features in terms of points, orientation, and scale from the input CT images having 1 × 96 length that is used for the detection of interest points as illustrated in Figure 5.
FIGURE 5

SURF features (a) original image (b) SURF features

SURF features (a) original image (b) SURF features

Histogram orientation gradient

The lesion regions have an irregular shape. Therefore, HOG (shape‐based) features are extracted from the segmented lesion region having a vector length is 1 × 8,100 and are shown in Figure 6.
FIGURE 6

Histogram orientation gradient (HOG) features (a) original image (b) HOG

Histogram orientation gradient (HOG) features (a) original image (b) HOG

Local binary pattern

The lung CT images having variant texture; therefore, LBP features are extracted from gray‐level images. The feature vector length is 1 × 59 and graphically shown in Figure 7.
FIGURE 7

Local binary patterns (LBP) of the original image

Local binary patterns (LBP) of the original image

SFTA

SFTA features are extracted from the segmented lesion region, which creates its own segment pattern and helps classify. The length of the feature vector is 1 × 22 and graphically shown in Figure 8.
FIGURE 8

SFTA features

SFTA features

DCNN features

Extracted deep features using pretrained deep learning models that is, ResNet 101 and inceptionv3. The ResNet101 contains 347 layers in which 01 input, 107 convolutional, 99 ReLU, 103 batch‐normalization, 01 maxpooling, 32 addition layers, 01 pool average global layer, fully connected, 01 softmax, & classification layer. While inceptionv3 contains 315 layers in which input, 96 convolutional, 94 batch‐normalization, 94 ReLU, 04 maxpooling, 09 average pooling layers, 14 depth‐concatenations, fully‐connected, classification& softmax layer. The fully connected layer named as FC1000 of Resnet101 and the prediction layer of inceptionv3 with 1 × 1000 features vector length is selected for classification and graphically plot in Figure 9.
FIGURE 9

DCNN features

DCNN features

Features selection and fusion

The features selection approach is presented using entropy as mentioned in Equation (2). The hand‐crafted and deep features are extracted from input images. The complete process of the feature's extraction, selection, and fusion is shown in Figure 10.
FIGURE 10

Features extraction, selection, and fusion process for COVID‐19 classification

Features extraction, selection, and fusion process for COVID‐19 classification The length of extracted features is such as fNHr1 × 128, fHOG1 × 8,100, fSFTA1 × 22, fSURF1 × 96, fResNet1011 × 1,000, and fInceptionv31 × 1,000. The best features are selected based on the maximum score using entropy. H denotes entropy, N represents the number of features, and n shows features at the same score level. The features are selected based on maximum score such as, 15 NHr features out of 128, 3,935 HOG features out of 8,100, 15 SFTA features out of 22, 90 SURF features out of 96, and 100 deep features out of 1,000 (ResNet101 and Inceptionv3). The selected score‐based features vectors are denoted as 1 × a, 1 × b, 1 × c, 1 × d, 1 × e, 1 × f. The selected features are concatenated/fused horizontally and denoted as FV1 × z . Where, and The fused features vector length is 1 × 4256 in which the last column represents the class labels. The final features vectors with class labels are passed to the ensemble classifiers.

Classification

The vector of resultant fused feature is fed to classifiers. The ensemble classifiers are used with three kernels such as a boosted tree (Freund, 2009), bagged tree (Breiman, 1996), and RUSBoosted tree (Seiffert, Khoshgoftaar, Van Hulse, & Napolitano, 2008). The ensemble classifiers are built by combining the week decision tree classifiers to create a powerful prediction model. Brief overview of the selected classifiers with the learning parameters is given in Table 3.
TABLE 3

Brief overview of ensemble classifiers

Model typeEnsemble approachLearnerMaximum splitsTotal learnersRate of the learningTraining time
Boosted TreeAdaBoostDecision tree20300.1380.3 s
Bagged TreeBag87764.689 s
RUSBoosted TreeRUSBoost20189.98 s
Brief overview of ensemble classifiers

BENCHMARK DATASETS

The proposed research is evaluated using three benchmark datasets. Dataset I, COVID‐19 segmentation data consists of the 40 patients, in which 100 slices of CT images with ground truth in NII format. Besides, this dataset out of 829 slices includes 373 positive cases (COVID‐19). Dataset II, consists of the Chinese Hospital CT imaging data of 216 patients. The patient's data contains 316 positive and 700 negative slices (Zhao, Zhang, He, & Xie, 2020). Dataset III includes 30 patients of POF Hospital, in which 2000 positive slices are included out of 4,500. The classified radiologists also created the ground truth. The description of training and testing images is given in Table 4.
TABLE 4

Training and testing data on benchmark datasets

DatasetTotal imagesTraining and testing imagesClassification
Dataset I829456 negative images and 373 positive images10 fold cross validation (means divide whole data into 10 sets, one randomly selected set is utilized for testing and remaining nine for training. Same methodology is repeated in all 10 iterations.
Dataset II1,016316 positive and 700 negative slices
Training and testing data on benchmark datasets

Experimentation

The proposed model is implemented on MATLAB core i7 CPU with Nvidia Graphic card. Two experiments are performed to test the proposed method performance. In Experiment 1, pixel‐based segmentation results are evaluated with ground truth. In Experiment 2, classifications are performed using hand‐crafted fused features and hand‐crafted and deep features fusion. In this experiment, training/testing data is divided on cross‐validation of 10‐fold. The performance of the proposed approach is assessed on various types of performance measurements described in Table 5.
TABLE 5

Performance measurements

MeasuresFormulas
IoU ϑϑ+ρ+δ here ϑ denotes true positive, false positive (ρ), and false negative (δ)
Accuracy ϑ+ϑ+ρ+δ+ ∃ represents the true negative
Sensitivity ϑϑ+δ
Specificity ρ+
Performance measurements

Experiment #1 (segmentation of lung region affected by COVID19)

In this experiment, segmented lesion region is evaluated with ground truth. IoU is mathematically expressed as: The segmentation model is trained on selected parameters, which are finalized after the extensive experiments alluded in Table 6.
TABLE 6

Experiment for the selection of optimal parameters

Encoder depthOptimizerEpochsLearning rateTraining accuracy
2Adam300.0020.96
3 Sgdm 40 0.001 0.98
3RmsProp1000.0010.95

Note: To prominant the selected parameters used and the results acheived from the proposed method are bold.

Experiment for the selection of optimal parameters Note: To prominant the selected parameters used and the results acheived from the proposed method are bold. Table 6, shows learning parameters of the proposed segmentation model where different values are selected for training of model and to analayze the accuracy. The different kinds of optimizers such as sgdm, adam, and rmsProp are employed with 40, 30, and 100 epochs, respectively. We observed that learning rate is 0.001, 40 epochs, and encoder depth 3 with sgdm optimizer provide greater training accuracy than other optimizers. These parameters are reflected in italic highlighted font and utilized for further experiments. The quantitative segmentation results are reflected in the Table 7.
TABLE 7

Segmentation outcomes on benchmark datasets

DatasetGlobal ACCMean ACCMean IoUWeighted IoUBF mean score
COVID‐19 segmentation0.980.880.870.960.89
POF hospital1.001.001.001.001.00
Segmentation outcomes on benchmark datasets Table 3 shows that the modified U‐Net segmentation model achieves 0.96 IoU on the COVID‐19 segmentation dataset and 1.00 IoU on the POF Hospital dataset. The segmentation results of the lung affected region with ground truth are shown in Figures 11 and 12.
FIGURE 11

Segmentation outcome on COVID‐19 segmentation dataset

FIGURE 12

Segmentation outcome on POF Hospital dataset

Segmentation outcome on COVID‐19 segmentation dataset Segmentation outcome on POF Hospital dataset

Experiment #2 (classification based on hand‐crafted fused features)

The classification experiment is performed by extraction of the different hand‐crafted features such as NHr, SFTA (Costa, Humpire‐Mamani, & Traina, 2012), HOG(Dalal & Triggs, 2005), LBP (Ojala, Pietikainen, & Maenpaa, 2002), and SURF (Bay, Ess, Tuytelaars, & Van Gool, 2008), and at the same time informative features are selected using entropy and fused serially (1 × 4055) vector length. That is provided to the ensemble method with three powerful learning models such as boosted, bagged, and the Rusboosted tree. The ensemble model prediction is graphically presented in the Figure 13.
FIGURE 13

Ensemble prediction model based on hand crafted features vector

Ensemble prediction model based on hand crafted features vector The quantitative classification outcomes are presented in Table 8, 9, 10.
TABLE 8

COVID‐19 classification on COVID‐19 segmentation dataset

Classification modelSensitivitySpecificityAccuracy
Boosted tree0.900.9186.3%
Bagged tree0.980.9796.8%
RUSBoosted tree1.00.9798.17%
TABLE 9

COVID‐19 classification on Chinese hospital dataset

Classification modelSensitivitySpecificityAccuracy
Boosted Tree0.970.9292.03%
Bagged Tree0.900.7089.52%
RUSBoosted Tree0.991.098.97%
TABLE 10

COVID‐19 classification on POF hospital dataset

Classification modelSensitivitySpecificityAccuracy
Boosted Tree1.01.099.77%
Bagged Tree0.960.9996.06%
RUSBoosted Tree0.990.9393.85%
COVID‐19 classification on COVID‐19 segmentation dataset COVID‐19 classification on Chinese hospital dataset COVID‐19 classification on POF hospital dataset The results in Tables 8, 9, 10 shows that RUSBoosted tree ensemble model provided better results in term of ACC, such as 98.17% on COVID‐19 segmentation dataset and 98.97% on Chinese Hospital dataset. At the same time, Boosted Tree ensemble model provides the highest accuracy on the POF hospital dataset.

Experiment 2.1 horizontally concatenation of the hand crafted and deep convolutional neural network

Extracted deep features are using Deep Convolutional Neural Network (DCNN) such as ResNET101(He, Zhang, Ren, & Sun, 2016) and Inception v3(Szegedy, Vanhoucke, Ioffe, Shlens, & Wojna, 2016) where length of the deep extracted features vector is 1 × 1,000. One hundred active features are selected from each feature vector using entropy and selected deep features. They are serially fused having a vector length 1 × 200 and further fused with hand‐crafted features vector length of 1 × 4,055. The length of final fused features is 1 × 4,255 which is passed to the ensemble learning classifiers for classification. The results with class labels are presented in Figure 14.
FIGURE 14

Confusion matrix using ensemble methods (a) bagged tree, (b) RUSBoosted tree, (c) Boosted tree

Confusion matrix using ensemble methods (a) bagged tree, (b) RUSBoosted tree, (c) Boosted tree In Figure 14, d denotes the disease class and h represents the healthy. The model prediction is performed using the combination of classical and deep features as illustrated in Figure 15.
FIGURE 15

An ensemble prediction model based on hand‐crafted and deep features fusion

An ensemble prediction model based on hand‐crafted and deep features fusion The empirical results are provided in Tables 11, 12, 13.
TABLE 11

COVID‐19 classification on COVID‐19 segmentation dataset

Classification modelSensitivitySpecificityAccuracy
Boosted Tree0.991.099.09%
Bagged Tree1.001.00100%
RUSBoosted Tree0.980.9796.8%
TABLE 12

COVID‐19 classification on Chinese hospital dataset

Classification modelSensitivitySpecificityAccuracy
Boosted Tree0.981.098.52%
Bagged Tree0.970.9292.14%
RUSBoosted Tree0.890.9788.61%
TABLE 13

COVID‐19 classification on POF hospital dataset

Classification modelSensitivitySpecificityAccuracy
Boosted Tree0.971.097.72%
Bagged Tree0.910.9689.73%
RUSBoosted Tree0.960.9291.25%
COVID‐19 classification on COVID‐19 segmentation dataset COVID‐19 classification on Chinese hospital dataset COVID‐19 classification on POF hospital dataset The results in Tables 11, 12, 13 show that the ensemble boosted tree, achieved 99.09% ACC on the COVID‐19 segmentation dataset, 98.52% ACC on the Chinese Hospital dataset, and 97.72% ACC on the POF Hospital dataset. While ensemble bagged tree obtained 100% ACC on segmentation of COVID‐19 dataset. The ratio among the AUC and ROC is also plotted in Figure 16.
FIGURE 16

AUC across ROC of ensemble methods (a) Bagged tree, (b) Boosted tree, (c) RUSBoosted tree

AUC across ROC of ensemble methods (a) Bagged tree, (b) Boosted tree, (c) RUSBoosted tree In Figure 16, the maximum achieved AUC is 1.00 using the bagged tree ensemble method, 0.95 AUC on the Boosted tree ensemble method, where 0.97 AUC on the RUSBoosted tree ensemble. The classification results are analyzed by ensembling different features such as texture, shape, NHr, and deep features. The findings of the proposed procedure are contrasted with the recent studies published alluded to in Table 14.
TABLE 14

Comparison of the proposed method with existing works

ReferencesDatasetsMethodResults accuracy%
Yang et al. (2020)UCSD‐AI4H/COVID‐CTResNet‐500.89
Burgos‐Artizzu (2020)Inception83.5 ± 2.0
Z. Wang, Liu, and Dou (2020)COVID‐Net78.69 ± 1.54
Ewen and Khan (2020)DenseNet1690.8621 ± 0.0474
Proposed methodFusion of hand crafted and deep features0.9852
Comparison of the proposed method with existing works The proposed framework is contrasted with four latest frameworks that is, (Burgos‐Artizzu, 2020; Ewen & Khan, 2020; Wang, Liu, et al., 2020; Yang et al., 2020). The existing published methods classified the COVID‐19 CT images using different pretrained CNN models, these methods achieved a maximum of 0.89 accuracy. However, this research provides a unique framework based on the deeper analysis of the deep and hand crafted features with more prominent features selection using entropy. The proposed approach provides a significant improvement of 0.9852 accuracy for the classification of COVID‐10 lung CT images. This approach might be utilized in medical domain to analyze the more informative features for disease classification. In this study, after comprehensive features analysis, it is observed that serially fusion of the hand‐crafted and deep features are provided a good effect on the model accuracy.

CONCLUSION

COVID‐19 segmentation and classification are an intricate task at an early stage. The accurate COVID‐19 segmentation is a crucial task because lesion regions appear at the boundaries that are difficult to segment using classical segmentation methods. Therefore, a deep learning model can be used to accurately segmentation the lesion region that appears at the borders. In this research, the modified U‐Net model is employed for segmentation. The proposed segmentation model achieved 0.98 global ACC, 0.96 weighted IoU on the COVID‐19 segmentation dataset, and 0.98 global ACC, 1.00 IoU on the POF Hospital dataset. The experimental discussion concludes that the fusion of the hand‐crafted and deep features provided competent results as compared to the fusion of hand‐crafted features only. The models achieved ACC of 99.09% on the Boosted Tree ensemble, 100% on the Bagged tree ensemble, and 96.8% RUSBoosted tree ensemble model.

CONFLICT OF INTEREST

All authors declared that there is no conflict of interest.

AUTHOR CONTRIBUTIONS

All authors are contributed equally in this manuscript.
  21 in total

1.  A Heuristic Neural Network Structure Relying on Fuzzy Logic for Images Scoring.

Authors:  Cheng Kang; Xiang Yu; Shui-Hua Wang; David S Guttery; Hari Mohan Pandey; Yingli Tian; Yu-Dong Zhang
Journal:  IEEE Trans Fuzzy Syst       Date:  2020-01-13       Impact factor: 12.029

2.  Clinical Characteristics of 138 Hospitalized Patients With 2019 Novel Coronavirus-Infected Pneumonia in Wuhan, China.

Authors:  Dawei Wang; Bo Hu; Chang Hu; Fangfang Zhu; Xing Liu; Jing Zhang; Binbin Wang; Hui Xiang; Zhenshun Cheng; Yong Xiong; Yan Zhao; Yirong Li; Xinghuan Wang; Zhiyong Peng
Journal:  JAMA       Date:  2020-03-17       Impact factor: 56.272

Review 3.  Fundus image classification methods for the detection of glaucoma: A review.

Authors:  Tanzila Saba; Syedia Tahseen Fatima Bokhari; Muhammad Sharif; Mussarat Yasmin; Mudassar Raza
Journal:  Microsc Res Tech       Date:  2018-10-03       Impact factor: 2.769

4.  Sensitivity of Chest CT for COVID-19: Comparison to RT-PCR.

Authors:  Yicheng Fang; Huangqi Zhang; Jicheng Xie; Minjie Lin; Lingjun Ying; Peipei Pang; Wenbin Ji
Journal:  Radiology       Date:  2020-02-19       Impact factor: 11.105

5.  Systematic review of artificial intelligence techniques in the detection and classification of COVID-19 medical images in terms of evaluation and benchmarking: Taxonomy analysis, challenges, future solutions and methodological aspects.

Authors:  O S Albahri; A A Zaidan; A S Albahri; B B Zaidan; Karrar Hameed Abdulkareem; Z T Al-Qaysi; A H Alamoodi; A M Aleesa; M A Chyad; R M Alesa; L C Kem; Muhammad Modi Lakulu; A B Ibrahim; Nazre Abdul Rashid
Journal:  J Infect Public Health       Date:  2020-07-01       Impact factor: 3.718

6.  First Case of 2019 Novel Coronavirus in the United States.

Authors:  Michelle L Holshue; Chas DeBolt; Scott Lindquist; Kathy H Lofy; John Wiesman; Hollianne Bruce; Christopher Spitters; Keith Ericson; Sara Wilkerson; Ahmet Tural; George Diaz; Amanda Cohn; LeAnne Fox; Anita Patel; Susan I Gerber; Lindsay Kim; Suxiang Tong; Xiaoyan Lu; Steve Lindstrom; Mark A Pallansch; William C Weldon; Holly M Biggs; Timothy M Uyeki; Satish K Pillai
Journal:  N Engl J Med       Date:  2020-01-31       Impact factor: 91.245

7.  Prediction of COVID-19 - Pneumonia based on Selected Deep Features and One Class Kernel Extreme Learning Machine.

Authors:  Muhammad Attique Khan; Seifedine Kadry; Yu-Dong Zhang; Tallha Akram; Muhammad Sharif; Amjad Rehman; Tanzila Saba
Journal:  Comput Electr Eng       Date:  2020-12-30       Impact factor: 3.818

8.  Viral reverse engineering using Artificial Intelligence and big data COVID-19 infection with Long Short-term Memory (LSTM).

Authors:  Ahmad M Abu Haimed; Tanzila Saba; Ayman Albasha; Amjad Rehman; Mahyar Kolivand
Journal:  Environ Technol Innov       Date:  2021-04-02

9.  Automatic Detection of Coronavirus Disease (COVID-19) in X-ray and CT Images: A Machine Learning Based Approach.

Authors:  Sara Hosseinzadeh Kassania; Peyman Hosseinzadeh Kassanib; Michal J Wesolowskic; Kevin A Schneidera; Ralph Detersa
Journal:  Biocybern Biomed Eng       Date:  2021-06-05       Impact factor: 4.314

10.  Deep Learning-Based COVID-19 Detection Using CT and X-Ray Images: Current Analytics and Comparisons.

Authors:  Amjad Rehman; Tanzila Saba; Usman Tariq; Noor Ayesha
Journal:  IT Prof       Date:  2021-06-18       Impact factor: 2.626

View more
  4 in total

1.  Liver Tumor Localization Based on YOLOv3 and 3D-Semantic Segmentation Using Deep Neural Networks.

Authors:  Javaria Amin; Muhammad Almas Anjum; Muhammad Sharif; Seifedine Kadry; Ahmed Nadeem; Sheikh F Ahmad
Journal:  Diagnostics (Basel)       Date:  2022-03-27

2.  An intelligence design for detection and classification of COVID19 using fusion of classical and convolutional neural network and improved microscopic features selection approach.

Authors:  Javaria Amin; Muhammad Almas Anjum; Muhammad Sharif; Tanzila Saba; Usman Tariq
Journal:  Microsc Res Tech       Date:  2021-05-08       Impact factor: 2.893

3.  COVID-DAI: A novel framework for COVID-19 detection and infection growth estimation using computed tomography images.

Authors:  Tahira Nazir; Marriam Nawaz; Ali Javed; Khalid Mahmood Malik; Abdul Khader Jilani Saudagar; Muhammad Badruddin Khan; Mozaherul Hoque Abul Hasanat; Abdullah AlTameem; Mohammad AlKathami
Journal:  Microsc Res Tech       Date:  2022-02-23       Impact factor: 2.893

4.  Recognition of Knee Osteoarthritis (KOA) Using YOLOv2 and Classification Based on Convolutional Neural Network.

Authors:  Usman Yunus; Javeria Amin; Muhammad Sharif; Mussarat Yasmin; Seifedine Kadry; Sujatha Krishnamoorthy
Journal:  Life (Basel)       Date:  2022-07-27
  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.