Literature DB >> 34898796

Local binary pattern and deep learning feature extraction fusion for COVID-19 detection on computed tomography images.

Auwalu Saleh Mubarak1, Sertan Serte1, Fadi Al-Turjman2, Zubaida Sa'id Ameen3, Mehmet Ozsoz3.   

Abstract

The deadly coronavirus virus (COVID-19) was confirmed as a pandemic by the World Health Organization (WHO) in December 2019. It is important to identify suspected patients as early as possible in order to control the spread of the virus, improve the efficacy of medical treatment, and, as a result, lower the mortality rate. The adopted method of detecting COVID-19 is the reverse-transcription polymerase chain reaction (RT-PCR), the process is affected by a scarcity of RT-PCR kits as well as its complexities. Medical imaging using machine learning and deep learning has proved to be one of the most efficient methods of detecting respiratory diseases, but to train machine learning features needs to be extracted manually, and in deep learning, efficiency is affected by deep learning architecture and low data. In this study, handcrafted local binary pattern (LBP) and automatic seven deep learning models extracted features were used to train support vector machines (SVM) and K-nearest neighbour (KNN) classifiers, to improve the performance of the classifier, a concatenated LBP and deep learning feature was proposed to train the KNN and SVM, based on the performance criteria, the models VGG-19 + LBP achieved the highest accuracy of 99.4%. The SVM and KNN classifiers trained on the hybrid feature outperform the state of the art model. This shows that the proposed feature can improve the performance of the classifiers in detecting COVID-19.
© 2021 John Wiley & Sons Ltd.

Entities:  

Keywords:  COVID‐19; LBP; deep learning; feature extraction; machine learning

Year:  2021        PMID: 34898796      PMCID: PMC8646483          DOI: 10.1111/exsy.12842

Source DB:  PubMed          Journal:  Expert Syst        ISSN: 0266-4720            Impact factor:   2.812


INTRODUCTION

A deadly respiratory disease was named Corona Virus 2 (SARS‐CoV‐2/COVID‐19) different from SARS‐CoV and MERS‐CoV was reported in Wuhan, Hubei Province, China (Lu et al., 2020) in few days the deadly disease spread to many cities in China and other countries and it was suspected that the disease spread was to have started from wild animals which are illegally sold in a seafood market in Wuhan (Zhu et al., 2020). On 30 January 2020, World Health Organization (WHO) declared COVID 19 as a global pandemic (Ye et al., 2020), the common symptoms of the disease include fever, dry cough, fatigue, lymphopenia and acute respiratory distress with severe pneumonia (Wang et al., 2020). Many people are being isolated for 14 days, the isolation serves as a preventive method of curving the spread of the virus, this isolation method is costly and time‐consuming as the patients present mild symptoms and there is no efficient method of detecting the virus at an early stage (Xie & Chen, 2020). The most widely used diagnosis of COVID‐19 infections is a reverse transcription‐polymerase chain reaction (RT‐PCR) examination of nasopharyngeal swabs (Chen et al., 2009). However, the high false‐negative rate (Chan et al., 2020; Chu et al., 2020) duration of the test and the lack of RT‐PCR test kits for the early stages of the epidemic can restrict the early diagnosis of infected patients. Computed tomography (CT) and chest X‐ray (CXR) is well matched to the lung picture of COVID‐19 infections in comparison to the swab examination. With the current situation in the world, a fast and reliable method of detecting COVID‐19 is needed, in profiling COVID‐19 X‐ray and CT scan images were used (Kroft et al., 2019; Liu et al., 2020; Strunk et al., 2014). CT and CXR show the spatial position of the alleged pathology as well as the extent of the damage. The signature phenotype of CXR is the bilateral spread of peripheral hazy lung opacities, including the consolidation of air space (Wong et al., 2020). The benefit of imaging is that it has a high sensitivity, a short response time, and can imagine the magnitude of the infection in the lung. The downside of imaging is that it has a low specificity, which makes it difficult to differentiate between different forms of lung infection, particularly when the lung infection is serious. Computer‐aided diagnostic (CAD) systems can help radiologists improve accuracy rate. Researchers are currently using handcrafted or learning features that are centred on lung texture, structure, and morphological characteristics for identification. However, it is always important and difficult to select the right classifier that can optimally handle the properties of the lung spaces. The traditional image recognition methods are support vector machine (SVM), K‐nearest neighbours (KNN), artificial neural networks (ANNs), decision trees (DTs) and Bayesian networks (BNs). These machine‐learning methods (Fehr et al., 2015; Orrù et al., 2012) needs hand‐crafted features to compute such as morphological, texture, SIFT, entropy, the density of pixels, elliptic shape, geometry, Fourier descriptors (EFDs), and off‐shelf classifiers as explained in Chen et al. (2021). In comparison, machine‐based learning (ML) approaches are known as non‐deep learning methods. There are many uses for these non‐deep learning approaches, such as the use of neurodegenerative diseases, cancer diagnosis, and psychological disorders (Cruz & Wishart, 2006; Doyle et al., 2007; Oakden‐Rayner et al., 2017; Orrù et al., 2012; Parmar et al., 2017). However, the major drawbacks of non‐deep learning approaches are that they rely on the extraction stage of the function and this makes it challenging to find the most important feature required to produce the most successful outcome. The application of artificial intelligence (AI) will be used to solve these difficulties. AI technology in the field of medical imaging is becoming increasingly popular, particularly for the advancement of technology and the growth of deep learning (Gao et al., 2017; Sa et al., 2021; Wang et al., 2017; Zhang et al., 2016). Convolutional neural networks (CNNs) have attained state‐of‐the‐art performance in the field of medical imaging based on previous studies (Waheed et al., 2020; Wang, Sun, et al., 2019; Wang, Tang, et al., 2019). This level of reliability is achieved by training and fine‐tuning the system's millions of parameters with labelled data. Because of the large number of parameters, CNN can easily overfit small datasets, so generalization efficiency is proportional to the size of the labelled data. Due to the limited number of datasets, limited datasets prove to be the most challenging problem in the medical imaging domain (Greenspan, 2016; Roth et al., 2015; Tajbakhsh et al., 2016). Medical image acquisition is a very expensive and tedious process that requires the participation of radiologists and researchers (Greenspan, 2016). Furthermore, due to the recent severity of the COVID‐19 disease, sufficient data of chest CT scan images are tough to obtain, unlike in Waheed et al. (2020) whereby the detection of COVID‐19 was performed on synthetic CT scan images, we proposed an offline data augmentation were by several data augmentations employed in many studies were performed on each of the three datasets employed in the study such as random reflection, random rotation, random rescale, random translation. Feature extraction is a critical component of a detection system's performance (Niu & Suen, 2012). CNN features are automatically trained. One of CNN's advantages in the case of transformations such as translation, scaling, and rotation is that they can be invariant. Invariance, rotation, and scale are three of CNN's most unique advantages, particularly in image recognition problems such as object detection, since they allow the network to abstract identity, enabling it to recognize the object even though the image's pixel values vary greatly. Feature extraction increases the accuracy of the models learned by extracting the features from the input data. This move in the general framework reduces data dimensionality by removing redundant data. It also increases the speed and inference of model training. Methods of extraction of features generate new features by rendering the variations and transformations of the original features. Colour, shape, texture or pixel value is the type of characteristics that can be obtained from medical images. Any diagnostic image, such as CT scan images, does not contain any colour detail. This is appreciated in this field. By linking most of the real objects around us to the internet, IoT technology helps us to integrate the physical and virtual environments. It gives everyday objects computational power and network access, allowing them to produce and disseminate data. These objects could include home appliances, wearable devices, medical equipment, and vehicles. The internet of things (IoT) pushes us closer to AI, smart access, and automation with less human interference (Alsukayti, 2020). Remote diagnostics, such as radiological services and online image processing, may benefit from the image database with advanced analysis, which allows physicians to diagnose critical illnesses without having to travel to remote locations. Improvements in biomedical and health‐care environments are apparent when combined with ML algorithms that can analyse and develop sophisticated simulations. Every day, millions of images are created, allowing different types of AI to open up new frontiers in big data analytics. The machine learning algorithms mainly clean structured data from raw datasets, converting it into expectations and assumptions to aid in the adoption of immediate actions (Xing et al., 2016). The IoT has spurred the development of a wide range of smart IoT applications across several industries. Successful IoT implementations and experiments are needed to advance the various technical aspects of these solutions. Low‐cost and modular approaches such as mathematical modelling and simulation are commonly used to solve this. However, such techniques are limited in their ability to realistically capture physical characteristics and network conditions. To address this issue, a revolutionary IoT testbed device that allows for the realistic testing of various IoT solutions in a controlled environment. The testbed was built to provide multidimensional general‐purpose support for various IoT properties such as sensing, connectivity, portal, energy storage, data processing, and security (Mu et al., 2019; Reshi et al., 2019; Schuß et al., n.d.). IoT and big data analytics, in general, are two main innovations that can change the biomedical and healthcare industries and improve people's lives (Banerjee et al., 2020).

Related works

The deadly disease of COVID‐19 has affected the world by crippling operations, various approaches have been taken into account in the battle against the spread of deadly disease, prediction analysis methods have been conducted to forecast the spread of disease, taking into account the number of infected, susceptible and recovered patients, and this prediction is a classic technique (Srivastava et al., 2020). to achieve real‐time prediction, AI models combined with IoT were used to help health professionals treat and monitor COVID‐19 by examining parameters such as temperature, blood pressure and heart rate, considering the high number of incidents, the protection of data transmission and the energy efficiency of the low‐power device used to gather information is very important, as suggested by Al‐turjman and Deebak (2020). To reduce the economic effect of COVID‐19 (Rahman, 2020) Proposed AI model that is data‐driven to forecast lock‐down and non‐lock‐down geographical borders to minimize the economic effects of the COVID‐19 pandemic, the accepted form of lock‐down by several countries was full lock‐down, this method is not beneficial to the economy. Employ transfer learning by training and comparing model output, more data has been created using image augmentation and conditional generative adversarial network (CGAN) (Loey et al., 2020). Transfer learning has been reported to be able to solve the problem of a few datasets, and both data generation will improve the efficiency of the model based on the ResNet‐50 performance criterion. With data augmentation, the ResNet‐50 achieved the highest accuracy of 82.1%, a sensitivity of 0.77, a specificity of 0.876 and precision of 0.849. Grey level scale zone matrix along with SVM was used to identify CT scan images of COVID19 and non‐COVID19 images, based on adjusting hyperparameters validation fold 2, 5, 10, and 10 fold to obtain the maximum accuracy (Barstugan et al., 2020). Explainable deep learning classification approach was proposed by Soares et al. (2020). To distinguish COVID‐19 and healthy individuals with CT scan images, the proposed model surpassed seven distinct AI models and reached an accuracy of 97.3, a sensitivity of 0.955, an F1‐score of 0.973, a precision of 0.991 and an AUC of 0.973. Different AI models were used to distinguish COVID‐19 and stable individuals‐ray and CT scan of which (81.5%–95.2%) and (95.4%–100%) were obtained in CT scans and X‐ray (Ahsan et al., 2020; Xie, 2020).

Contributions

In this study, KNN and SVM classifiers were trained using handcrafted LBP extracted features and seven pre‐trained deep learning (i.e., CNN) models extracted features for COVID‐19 detection. The contribution in this study are: We proposed a hybrid LBP and CNN feature to train KNN and SVM classifiers. We compared the performance of KNN and SVM classifiers on handcrafted LBP and automatic CNN features. The performance of the KNN and SVM in detecting COVID‐19 was improved by concatenating the LBP and the CNN features. Three datasets were merged to generalized the performance of the model and improve the training of the model. The proposed model outperformed the state of the art model.

COVID‐19 DETECTION

This section describes the characteristics of the dataset used, the proposed feature extractions techniques and the machine learning models for COVID‐19 detection.

Dataset

In this study, three datasets were merged to efficiently classify COVID‐19, common pneumonia and healthy individual as presented in Table 1 (Figure 1).
TABLE 1

Compiled dataset

DatasetCOVID‐19 positiveCommon pneumoniaHealthy individuals
Yang et al. (n.d.)349NA397
Yan et al. (n.d.)371328NA
Soares et al. (2020)1252NA1229
Total number of CT scan images per class19723281608
FIGURE 1

Computed tomography (CT) scan images samples (a) COVID‐19 positive, (b) COVID‐19 healthy individuals, (c) common pneumonia

Compiled dataset Computed tomography (CT) scan images samples (a) COVID‐19 positive, (b) COVID‐19 healthy individuals, (c) common pneumonia

Support vector machines

The SVM (Wang et al., 2013) classification is referred to as a mechanism wherein the supervised binary classification system is used and where a classification model is implemented, in which the algorithm creates a hyperplane that maximizes the margin that occurs between two input classes. For instance, considering linearly separate data with two distinct classes, the system can have numerous hyperplanes which separate two classes. SMV considers the most optimal hyperplane with the highest margin of all available hyperplanes, where the margin is the difference between the hyperplane and the support vectors. Given a set of training data is the actual value, represents the input vector and is the data number), given that the SVM function is: where is mapped non‐linearly from input vector , which are input feature spaces. Then, the SVM equation is given as (Wang et al., 2013): is the kernel function in the feature space after performing non‐linear mapping and is bias term. The most commonly used kernel function is Gaussian Radial Basis Function (RBF) because it performs better than linear and polynomial kernel as it is not only capable to map non‐linearly training data into infinite‐dimensional space but also easier to implement (Wang et al., 2013) and it is given as: where is the kernel parameter.

K‐Nearest neighbours

KNN is a non‐parametric method of classification (Altman, 1992) when classifying using KNN, the entity to be ranked is decided upon by its neighbours and assigned to the most comparable class of its closest neighbours. Three‐class KNNs is used as part of the research.

Local binary pattern

The local binary pattern (LBP) is an efficient non‐parametric operator for the description of local image features and has provided the centre pixel (x c,y c), the ordered binary set identified as LBP is obtained by comparing the grey value of the centre pixel (x c,y c), with the pixels of its eight neighbours. The LBP code is thus expressed as the decimalized version of an octet binary integer: where i c represents the grey value of the middle pixel (x c,y c), and in the grey value of the pixels of its eight neighbours. The LBP code is invariant for any monotonous transformation of the grey level, and the local binary code remains unchanged after transformation.

Convolutional neural network

A typical CNN is a type of deep model in which convolutional filters and pooling operations are applied alternately on the local neighbourhoods of each pixel in the raw input, generating complex high‐level features (Li et al., 2017). CNN's have been primarily applied to 2D images, and have achieved better performance in image classification.

Transfer learning

Transfer learning is a research topic of machine learning. It focuses on preserving information learned when solving a particular problem and adapting it to a particular but connected problem (Apostolopoulos & Mpesiana, 2020; Taresh, 2020; Hussein et al., 2019; Learning, 2020; Mahmud et al., 2020). In training the pre‐trained network for another problem, some features of the pre‐trained models may be modified, such modifications are layers to freeze, layers to be inserted, and some hyperparameter values changed.

ResNet

Residual network (ResNet) (He et al., 2016). It is a deep learning algorithm used to classify images. The core principle behind ResNet is to deal with vanishing gradients that degrade network performance caused by accumulating a convolution layer over a pooling layer in deep network architecture, shortcuts that provide identification is a residual block, the notion of adding skip connections essentially eliminates a high training error, other deep networks do not include an identity link, which is why ResNet may not. The input layer accepts a 224 × 224 image size.

GoogleNet

GoogleNet is a 22‐layer network composed of the input layer, convolution layers, max‐pooling and softmax classifier, the key feature that makes GoogleNet different is the 1 × 1 convolution, networking and global average pooling. GoogleNet won the 2014 ILSVRC competition at a low error rate relative to VGG (Szegedy et al., 2015).

ShuffleNet

ShuffleNet (Zhang et al., n.d.) is a 50‐layer deep learning model that uses AlexNet's group convolution on the first convolution layer. Although group convolution greatly decreases computation, one downside is that the performance of such channels is powered by a small fraction of the input. To resolve this problem, to address this issue, the channels which are also differentiable are shuffled in ShuffleNet to address this issue.

VGG

VGG (Karen & Andrew, 2015) VGG16 has 16 layers, 5 convolutional layers, 3 trainable layers, and the remaining layers are max‐pooling layers, while VGG19 has 19 layers. This design came in second place in the 2014 Visual Recognition Competition, that is, ILSVRC‐2014.

TRAINING

in this study three dataset were merged to detect COVID‐19, the three datasets contained three classes, COVID‐19, common pneumonia and healthy individuals. Before training, the data was preprocessed by resizing the images to 224 by 224, also, several data augmentation such as random reflection, random rotation, random rescale, random translation along X‐axis and random translation along Y‐axis was performed to increase the number of images, improves training and to reduce overfitting (Ghassemi et al., 2020; Loey et al., 2020). After the augmentation, two features extraction techniques handcrafted LBP and automatic deep features from seven pre‐trained models MobileNetv2, GoogleNet, ResNet‐50, ResNet‐101, ShuffleNet, VGG16 and VGG19 were extracted. Eighty percentage of the data was used for training and 20% for testing. The training for this study is carried out in stages using two classifiers, KNN and SVM, in the first stage, textual features were extracted using LBP and were classified by KNN and SVM, in the second stage, high‐level deep features were extracted from each of the seven pre‐trained models MobileNetv2, GoogleNet, ResNet‐50, ResNet‐101, ShuffleNet, VGG16 and VGG19, the features of the last pooling layer of each network were used for the training of the two classifiers KNN and SVM, in the third stage the LBP features and features extracted from each of the seven pre‐trained networks are concatenated then classified using the two classifiers KNN and SVM, the concatenation of the two features will give the advantage of textual and high‐level deep features of the CT scan images. Figure 2 shows the classification process for the COVID‐19 detection.
FIGURE 2

Shows the classification process for the COVID‐19 detection

Shows the classification process for the COVID‐19 detection

RESULTS AND DISCUSSION

Within the scope of the study, three classes of CT scan images COVID‐19, common pneumonia and healthy individuals classes were classified using two classifiers, KNN and SVM. Before training the classifiers, handcrafted LBP features and deep high‐level features of seven deep learning models were extracted. In Table 2, SVM and KNN classifiers performance were compared to determine the best performing model with LBP features as inputs, the SVM achieves an accuracy of 97.5% sensitivity of 98.7% a specificity of 96.1%, F1 score of 97.5%, the precision of 96.4%, Yonden index of 94.8% and AUC of 97.4%. The SVM as shown in Figure 3 outperformed the KNN in terms of accuracy, sensitivity, specificity, F1 score, precision, Yonden index and AUC. This shows that the SVM can efficiently detect COVID‐19, common pneumonia and healthy individuals CT scan images based on LBP features extracted from the CT scan images.
TABLE 2

SVM and KNN performance on LBP extracted features

ModelsAccuracy (%)Sensitivity (%)Specificity (%)F1 score (%)Precision (%)Yonden index (%)AUC (%)
LBP SVM97.598.796.197.596.494.897.4
LBP KNN91.892.590.992.091.583.491.7

Abbreviations: KNN, K‐nearest neighbour; LBP, local binary pattern; SVM, support vector machines.

FIGURE 3

Performance comparison between support vector machines (SVM) and K‐nearest neighbour (KNN) on local binary pattern (LBP) extracted features

SVM and KNN performance on LBP extracted features Abbreviations: KNN, K‐nearest neighbour; LBP, local binary pattern; SVM, support vector machines. Performance comparison between support vector machines (SVM) and K‐nearest neighbour (KNN) on local binary pattern (LBP) extracted features In Tables 3 and 4, SVM and KNN were used to classify seven deep learning models extracted features respectively, in Table 3, SVM with extracted features of VGG‐19 achieves an accuracy of 98.7, the sensitivity of 98.8%, specificity of 98.7%, F1 score of 98.8%, the precision of 97.5%, Yonden index of 97.5% and AUC of 98.7%. In Table 3, the SVM turns out to achieve the highest performance with features extracted from VGG‐19. The KNN classifier achieved the highest performance with features extracted from VGG‐16, as presented in Table 4, the KNN with VGG‐16 extracted features achieves an accuracy of 96.2%, the sensitivity of 96.9%, a specificity of 95.5%, F1 score of 96.3%, precision of 95.7%, Yonden Index of 92.4% and AUC of 96.2%. If we compare the performances of the SVM and KNN on both the LBP and the deep learning features, it will be observed in Figures 4 and 5 that the SVM shows superiority in terms of classifying the CT scan images for the three classes.
TABLE 3

CNN extracted features with SVM classifier models performance

ModelsAccuracy (%)Sensitivity (%)Specificity (%)F1 score (%)Precision (%)Yonden index (%)AUC (%)
ResNet‐5095.995.695.195.996.390.795.4
ResNet‐10195.995.696.195.996.391.795.9
VGG‐1697.597.497.597.597.594.997.5
VGG‐1998.798.898.798.898.897.598.7
GoogLeNet97.898.197.498.498.895.597.8
MobileNetv297.897.59897.898.1495.597.7
ShuffleNet93.48997.493.097.386.493.2

Abbreviations: CNN, convolutional neural network; SVM, support vector machines.

TABLE 4

CNN extracted features with KNN classifier models performance

ModelsAccuracy (%)Sensitivity (%)Specificity (%)F1 score (%)Precision (%)Yonden index (%)AUC (%)
ResNet‐5094.392.596.194.396.288.694.3
ResNet‐10194.996.393.595.194.089.894.9
VGG‐1696.296.995.596.395.792.496.2
VGG‐1993.1959093.091.28592.5
GoogLeNet90.891.389.390.890.280.690.3
MobileNetv293.195.690.393.391.285.993.0
ShuffleNet90.991.998.791.190.390.695.3

Abbreviations: CNN, convolutional neural network; KNN, K‐nearest neighbour.

FIGURE 4

Convolutional neural network (CNN) extracted features with support vector machines (SVM) classifier models performance

FIGURE 5

Convolutional neural network (CNN) extracted features with K‐nearest neighbour (KNN) classifier models performance

CNN extracted features with SVM classifier models performance Abbreviations: CNN, convolutional neural network; SVM, support vector machines. CNN extracted features with KNN classifier models performance Abbreviations: CNN, convolutional neural network; KNN, K‐nearest neighbour. Convolutional neural network (CNN) extracted features with support vector machines (SVM) classifier models performance Convolutional neural network (CNN) extracted features with K‐nearest neighbour (KNN) classifier models performance To take the advantage of both the handcrafted LBP features and the automatic deep learning features, in this study, the LBP and CNN features were extracted and then concatenated, the KNN and the SVM classifiers were trained using these concatenated features. In Tables 5 and 6, the results of the SVM and KNN classifiers were presented respectively. In Table 5, the SVM achieved the highest performance in terms of accuracy with ResNet‐50+ LBP and MobileNetv2+ LBP features with an accuracy of 98.7%, in terms of sensitivity VGG‐16+ LBP and MobileNetv2+ LBP achieved a sensitivity of 100% each, for the specificity ResNet‐50+ LBP achieves the highest specificity of 100%. In Table 6, the KNN classifier trained with VGG‐19 + LBP features shows superiority in terms of performance as it achieves an accuracy of 99.4%, a sensitivity of 99.3%, a specificity of 99.3%, F1 score of 99%, precision of 98.8%, Yonden Index of 98.6% and AUC of 99.3. Based on Tables 5 and 6, the concatenated features have improved the detection of COVID‐19 on CT scan images, as clearly shown in Figures 6 and 7, the two classifiers can effectively classify the new proposed concatenated features. In Table 7, compared to the state of the art models in which three classes of COVID‐19, common pneumonia and Healthy individuals CT scan images were detected, our models show greater performance in terms of accuracy, sensitivity specificity and AUC.
TABLE 5

CNN and LBP extracted features with SVM classifier models performance

ModelsAccuracy (%)Sensitivity (%)Specificity (%)F1 score (%)Precision (%)Yonden index (%)AUC (%)
ResNet‐50+ LBP98.797.510098.198.897.598.8
ResNet‐101+ LBP94.392.596.194.396.288.694.3
VGG‐16+ LBP97.510094.897.695.394.897.4
VGG‐19+ LBP98.198.797.498.197.696.198.05
GoogLeNet+ LBP98.197.598.798.198.896.298.1
MobileNetv2+ LBP98.710097.498.897.697.498.7
ShuffleNet+ LBP96.293.89793.292.790.895.4

Abbreviations: CNN, convolutional neural network; LBP, local binary pattern; SVM, support vector machines.

TABLE 6

CNN and LBP extracted features with KNN classifier models performance

ModelsAccuracy (%)Sensitivity (%)Specificity (%)F1 score (%)Precision (%)Yonden index (%)AUC (%)
ResNet‐50+ LBP99.199.398.798.798.29899
ResNet‐101+ LBP97.298.795.597.295.894.297.1
VGG‐16 + LBP98.197.598.798.198.896.298.1
VGG‐19 + LBP99.499.399.399.098.898.699.3
GoogLeNet + LBP98.498.79898.498.296.798.4
MobileNetv2+ LBP97.298.795.597.295.894.297.1
ShuffleNet+ LBP95.998.593.196.294.191.695.8

Abbreviations: CNN, convolutional neural network; KNN, K‐nearest neighbour; LBP, local binary pattern.

FIGURE 6

Convolutional neural network (CNN) and local binary pattern (LBP) extracted features with support vector machines (SVM) classifier models performance

FIGURE 7

Convolutional neural network (CNN) and local binary pattern (LBP) extracted features with K‐nearest neighbour (KNN) classifier models performance

TABLE 7

State of the art model comparison

Ref.ModelsAccuracy (%)SensitivitySpecificityAUC
Amyar et al. (2020)T1 & T2 & T3 512 × 51291.130.940.850.94
T1 & T2 & T3 256 × 25694.670.960.920.97
CNN and LBP extracted features with SVM classifier models performance Abbreviations: CNN, convolutional neural network; LBP, local binary pattern; SVM, support vector machines. CNN and LBP extracted features with KNN classifier models performance Abbreviations: CNN, convolutional neural network; KNN, K‐nearest neighbour; LBP, local binary pattern. Convolutional neural network (CNN) and local binary pattern (LBP) extracted features with support vector machines (SVM) classifier models performance Convolutional neural network (CNN) and local binary pattern (LBP) extracted features with K‐nearest neighbour (KNN) classifier models performance State of the art model comparison It is important to improve COVID‐19 detection, particularly on CT scan images that provide slice‐level information of the chest. X‐ray images were used in many experiments on medical imaging, but CT images were used in only a handful of studies. Hence, we were motivated to carry out this study based on several factors, including insufficient CT scan images data, and the ability to distinguish between common pneumonia, COVID‐19 and healthy individuals' CT scan images.

Performance criteria

The performance criteria of a model describe how well the model performs in solving a problem. The performance of the two pre‐trained models based on the binary and multi‐class classification was compared using the frequently applied performance evaluation criteria, namely validation accuracy (ACC), sensitivity (SN), specificity (SP), F1 score, precision (PR), Yonden‐index and AUC.

Accuracy

Accuracy is a measure that gives an insight into how well the model learns and produces reliable results. It is the proportion of predictions that were provided correctly by the model. The accuracy of a model is the ratio of correctly predicted samples to the number of input samples. The number of correctly predicted samples is the sum of the number of true positives and false negatives

Sensitivity or recall

It is defined as the ability of a model to test correctly and identify patients with a disease as presented in Equation (5).

Specificity

Specificity is a measure of how many negatives the trained model managed to capture out of the entire set of correctly predicted negative values by labelling the samples as negative. The relation for calculating specificity is presented in Equation (6).

F1‐Score

F1‐score is a measure of the balance between the precision and recall of a model. It is used to perform a statistical analysis of the test accuracy. The F1‐score of a model lies between 0 and 1. It is said to be very good if its value lies near 1 and very bad if it is near 0. It is calculated by applying Equation (7)

Yonden index

Is the cut‐point that optimizes the biomarker's distinguishing ability when equal weight is given to sensitivity and specificity. It also gives a summary of the receiver operating characteristic curve, as presented in Equation (8).

Precision

Precision is a measure of how precise or accurate the model is in terms of positive classifications. In other words, it measures the number of true positives out of all the predicted positives. The relation for precision is presented in Equation (9) where TP, true positive; TN, true negative; FP, false positive; FN, false negative.

CONCLUSION

In this study, two classifiers SVM and KNN were employed to classify COVID‐19, common pneumonia and healthy individuals CT scan images, before training the classifiers, handcrafted LBP features and Automatic deep learning features of seven pre‐trained networks were extracted, the training of the classifiers was conducted on the extracted features, to improve the performance of the classifiers, a new feature was proposed by concatenating the LBP and the CNN features to train the classifiers, this proposed feature shows improvement in the performance of the classifiers compared to the performance of the classifiers on LBP or CNN features. Due to the limited number of CT scan images available at this stage in the pandemic, only a small number of CT scan images were used for training the classifiers. This shows the scarcity of data in the research community. In the future, the authors will explore more pre‐processing techniques while more classifiers such as Decision Tree and ensemble classifiers will be explored to improve the performance of the detection. The authors will also try to incorporate the proposed model with IoT for easy detection of COVID‐19 around the globe.

CONFLICT OF INTEREST

The authors declare no conflict of interest.

AUTHOR CONTRIBUTIONS

Auwalu Saleh Mubarak, Sertan Serte, Fadi Al‐Turjman, Zubaida Sa'id Ameen and Mehmet Ozsoz contributed to the design and implementation of the research, to the analysis of the results and the writing of the manuscript.
  35 in total

1.  Can Chest CT Features Distinguish Patients With Negative From Those With Positive Initial RT-PCR Results for Coronavirus Disease (COVID-19)?

Authors:  Dandan Chen; Xinqing Jiang; Yong Hong; Zhihui Wen; Shuquan Wei; Guangming Peng; Xinhua Wei
Journal:  AJR Am J Roentgenol       Date:  2020-05-05       Impact factor: 3.959

2.  Convolutional Neural Networks for Medical Image Analysis: Full Training or Fine Tuning?

Authors:  Nima Tajbakhsh; Jae Y Shin; Suryakanth R Gurudu; R Todd Hurst; Christopher B Kendall; Michael B Gotway
Journal:  IEEE Trans Med Imaging       Date:  2016-03-07       Impact factor: 10.048

3.  Molecular Diagnosis of a Novel Coronavirus (2019-nCoV) Causing an Outbreak of Pneumonia.

Authors:  Daniel K W Chu; Yang Pan; Samuel M S Cheng; Kenrie P Y Hui; Pavithra Krishnan; Yingzhi Liu; Daisy Y M Ng; Carrie K C Wan; Peng Yang; Quanyi Wang; Malik Peiris; Leo L M Poon
Journal:  Clin Chem       Date:  2020-04-01       Impact factor: 8.327

4.  Automatic classification of prostate cancer Gleason scores from multiparametric magnetic resonance images.

Authors:  Duc Fehr; Harini Veeraraghavan; Andreas Wibmer; Tatsuo Gondo; Kazuhiro Matsumoto; Herbert Alberto Vargas; Evis Sala; Hedvig Hricak; Joseph O Deasy
Journal:  Proc Natl Acad Sci U S A       Date:  2015-11-02       Impact factor: 11.205

5.  Improving Computer-Aided Detection Using Convolutional Neural Networks and Random View Aggregation.

Authors:  Holger R Roth; Le Lu; Jiamin Liu; Jianhua Yao; Ari Seff; Kevin Cherry; Lauren Kim; Ronald M Summers
Journal:  IEEE Trans Med Imaging       Date:  2015-09-28       Impact factor: 10.048

6.  Multi-task deep learning based CT imaging analysis for COVID-19 pneumonia: Classification and segmentation.

Authors:  Amine Amyar; Romain Modzelewski; Hua Li; Su Ruan
Journal:  Comput Biol Med       Date:  2020-10-08       Impact factor: 4.589

7.  Momentum contrastive learning for few-shot COVID-19 diagnosis from chest CT images.

Authors:  Xiaocong Chen; Lina Yao; Tao Zhou; Jinming Dong; Yu Zhang
Journal:  Pattern Recognit       Date:  2021-01-16       Impact factor: 7.740

8.  A Novel Coronavirus from Patients with Pneumonia in China, 2019.

Authors:  Na Zhu; Dingyu Zhang; Wenling Wang; Xingwang Li; Bo Yang; Jingdong Song; Xiang Zhao; Baoying Huang; Weifeng Shi; Roujian Lu; Peihua Niu; Faxian Zhan; Xuejun Ma; Dayan Wang; Wenbo Xu; Guizhen Wu; George F Gao; Wenjie Tan
Journal:  N Engl J Med       Date:  2020-01-24       Impact factor: 91.245

9.  Covid-19: automatic detection from X-ray images utilizing transfer learning with convolutional neural networks.

Authors:  Ioannis D Apostolopoulos; Tzani A Mpesiana
Journal:  Phys Eng Sci Med       Date:  2020-04-03

10.  Chest CT for Typical Coronavirus Disease 2019 (COVID-19) Pneumonia: Relationship to Negative RT-PCR Testing.

Authors:  Xingzhi Xie; Zheng Zhong; Wei Zhao; Chao Zheng; Fei Wang; Jun Liu
Journal:  Radiology       Date:  2020-02-12       Impact factor: 11.105

View more
  1 in total

1.  Local binary pattern and deep learning feature extraction fusion for COVID-19 detection on computed tomography images.

Authors:  Auwalu Saleh Mubarak; Sertan Serte; Fadi Al-Turjman; Zubaida Sa'id Ameen; Mehmet Ozsoz
Journal:  Expert Syst       Date:  2021-09-29       Impact factor: 2.812

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.