Literature DB >> 35180500

QCovSML: A reliable COVID-19 detection system using CBC biomarkers by a stacking machine learning model.

Tawsifur Rahman1, Amith Khandakar1, Farhan Fuad Abir2, Md Ahasan Atick Faisal2, Md Shafayet Hossain3, Kanchon Kanti Podder4, Tariq O Abbas5, Mohammed Fasihul Alam6, Saad Bin Kashem7, Mohammad Tariqul Islam3, Susu M Zughaier8, Muhammad E H Chowdhury9.   

Abstract

The reverse transcription-polymerase chain reaction (RT-PCR) test is considered the current gold standard for the detection of coronavirus disease (COVID-19), although it suffers from some shortcomings, namely comparatively longer turnaround time, higher false-negative rates around 20-25%, and higher cost equipment. Therefore, finding an efficient, robust, accurate, and widely available, and accessible alternative to RT-PCR for COVID-19 diagnosis is a matter of utmost importance. This study proposes a complete blood count (CBC) biomarkers-based COVID-19 detection system using a stacking machine learning (SML) model, which could be a fast and less expensive alternative. This study used seven different publicly available datasets, where the largest one consisting of fifteen CBC biomarkers collected from 1624 patients (52% COVID-19 positive) admitted at San Raphael Hospital, Italy from February to May 2020 was used to train and validate the proposed model. White blood cell count, monocytes (%), lymphocyte (%), and age parameters collected from the patients during hospital admission were found to be important biomarkers for COVID-19 disease prediction using five different feature selection techniques. Our stacking model produced the best performance with weighted precision, sensitivity, specificity, overall accuracy, and F1-score of 91.44%, 91.44%, 91.44%, 91.45%, and 91.45%, respectively. The stacking machine learning model improved the performance in comparison to other state-of-the-art machine learning classifiers. Finally, a nomogram-based scoring system (QCovSML) was constructed using this stacking approach to predict the COVID-19 patients. The cut-off value of the QCovSML system for classifying COVID-19 and Non-COVID patients was 4.8. Six datasets from three different countries were used to externally validate the proposed model to evaluate its generalizability and robustness. The nomogram demonstrated good calibration and discrimination with the area under the curve (AUC) of 0.961 for the internal cohort and average AUC of 0.967 for all external validation cohort, respectively. The external validation shows an average weighted precision, sensitivity, F1-score, specificity, and overall accuracy of 92.02%, 95.59%, 93.73%, 90.54%, and 93.34%, respectively.
Copyright © 2022 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  COVID-19; Complete blood count (CBC); Detection; RT-PCR; Stacking machine learning

Year:  2022        PMID: 35180500      PMCID: PMC8839805          DOI: 10.1016/j.compbiomed.2022.105284

Source DB:  PubMed          Journal:  Comput Biol Med        ISSN: 0010-4825            Impact factor:   4.589


Introduction

After only one year and a half years’ duration of the pandemic, the SARS-CoV2 coronavirus, the pathogen responsible for COVID-19, has caused about 360 million infections with over five million casualties around the globe [1]. It is the worst pandemic that afflicted humanity since the Spanish flu of 1918, which overwhelmed the global healthcare systems and caused a worldwide economic crisis. Unlike other coronaviruses (e.g., SARS, MERS), SARS-CoV-2 can infect an individual without causing any symptoms or very mild and non-characteristic ones for a prolonged duration. This increases the risk of spreading the disease and early identification of patients becomes crucial [2]. To detect SARS-CoV-2 infections, the current gold standard tool is the reverse transcription-polymerase chain reaction (RT-PCR). However, it requires specialized equipment and reagents (at least 4–5 h under optimal conditions), trained personnel for the samples collection, and proper genetic conservation of the Ribonucleic acid (RNA) sequences used for annealing the primers [3]. The data science community has proposed several machine learning (ML) based methods to assist the diagnosis and contain the transmission of COVID-19 [4]. In several studies, a number of biomarkers have been found to identify COVID-19 patients at the risk of severe infection and death by providing insight into their underlying health conditions. Zheng et al. [5] found that neutrophil, platelet, and white blood cell counts were typically normal in the COVID-19 patients at the time of admission (87.9%, 85.1%, and 88.7%, respectively). However, the majority of the patients developed lymphopenia right after the onset of symptoms, and it became more pronounced with the progression of the disease. They also showed that the neutrophil, lymphocyte, and platelet (NLP) counts are clinically useful in stratifying patients. However, the dataset was small, and they did not test the model on any external dataset. Moreover, Wang et al. [6] used 301 adult patient data to develop a new prediction score called ANDC for evaluating the mortality risk of COVID-19 patients. LASSO regression was used to identify age, D-dimer, neutrophil-to-lymphocyte ratio (NLR), and C-reactive protein as death predictors for COVID-19 patients [6]. Although several clinical studies [5,7] have shown blood test-based diagnosis as an effective and low-cost solution for the early COVID-19 detection, only a few ML models have been applied to hematological parameters [4,6,[8], [9], [10], [11]]. Weng et al. [12] developed a nomogram using a logistic regression classifier on a single dataset collected from a hospital in Wuhan. Ramachandran et al. [8] used multivariable logistic regression analyses on a single dataset collected from a tertiary care academic medical center in New York City and found the deviations in the Red Blood Cell distribution width (RDW) among the hospitalized COVID-19 patients. A similar study was done by Foy et al. [10] using a dataset collected from Massachusetts General Hospital in Boston, Massachusetts, USA. Gong et al. [9] developed a nomogram using logistic regression on a dataset collected from Wuhan. A different study has developed several machine learning (ML) models to improve diagnostic capabilities to prevent the spread of the pandemic [4]. The majority of these models [4,[13], [14], [15], [16]] employed computed tomography (CT) scans and chest X-rays. However, the CT scan-based diagnostic approaches showed higher false-negative results that caused concern for real-life applications [17]. Moreover, CT imaging solutions are expensive, time-consuming, and require specialized equipment; hence, they are not feasible for frequent testing. To overcome the above-mentioned limitations, we develop a stacking machine learning approach to identify COVID-19 patients using blood-test biomarkers that are widely used in clinical practice and take a few minutes (under emergency conditions) and are much cheaper than RT-PCR test, chest X-ray, and CT imaging. Furthermore, we develop a nomogram-based scoring technique using the machine learning approach and the scoring technique uses complete blood count (CBC) parameters to diagnose COVID-19 patients in resource-constrained situations and countries, where the availability and application of RT-PCR are limited. Moreover, the developed method can be utilized in combination with the RT-PCR to improve its sensitivity. Different biomarkers have been used in several recent works on machine learning-based early mortality prediction systems [[18], [19], [20], [21], [22]]. However, to the best of the authors' knowledge, there is no prior work employing such biomarkers to diagnose COVID-19 patients that are trained with a large dataset and validated externally on datasets from different countries. Fast screening of COVID-19 with high sensitivity is crucial for both resource and treatment planning [19,20,[23], [24], [25], [26]]. Therefore, this study proposes a reliable COVID-19 diagnostic technique that can be generalized in other health care settings. It adds to the body of knowledge for developing a framework using a stacking machine learning technique and validates the performance with six completely unseen datasets. The remaining article is organized as stated: Section II explains the detailed methodology of the study along with the datasets used for this study, the data preprocessing stages, and the nomogram-based scoring technique, while Section III shows the results. Moreover, Section IV discusses the results and validates the performance of the nomogram-based scoring technique. Finally, the article is concluded in Section V discussing the future implications of this study.

Methodology

The study has used seven publicly available clinical datasets [27] for model training, internal and external validation. Among them, one was used for training and internal validation whilst others were used for external validation. Feature Engineering was done extensively on getting the best-trained model which involved pre-processing the data, using different feature ranking techniques, applying different machine learning classifiers with the help of top-ranked features. We proposed a stacking-based ML model leveraging the top-performing classifiers and performance compared with other conventional ML classifiers to identify COVID-19 positive patients. The trained model is used to develop a nomogram-based scoring system (Qu Covid Stacking-based ML (QCovSML)) system, which is externally validated with six clinical datasets collected from different hospitals and countries to prove/validate its robustness. Fig. 1 illustrates the schematic overview of the methodology.
Fig. 1

Step-by-step overview of the methodology.

Step-by-step overview of the methodology.

Database description/study population

In this study, we used seven different hospital clinical biomarker datasets [27] from three different countries: Italy, Brazil, and Ethiopia. While the largest dataset from Italy was kept for training, the remaining datasets were used for external validation. The training dataset had a total of 1624 instances and 21 features, collected upon admission at the emergency departments (ED) of two hospitals - IRCCS Hospital San Raaele (OSR) and the IRCCS Istituto Ortopedico Galeazzi (IOG), which were largely impacted during the first COVID19 pandemic in Milan, Italy. The dataset spanned from March 5, 2020 to May 26, 2020, and it was sufficiently balanced and heterogeneous. We have validated the developed model with the following six external datasets. This dataset contains 337 instances (163 positives, 174 negative) collected in March/April at the Desio Hospital. It was collected at the “Papa Giovanni XXIII” Hospital of Bergamo, containing 249 patients' data with 104 COVID-19 positive, and 145 COVID-19 negative patients. The dataset was collected between March and April of 2020. This dataset includes 224 patients' data with 118 COVID-19 positive and 106 COVID-19 negative patients collected in November 2020 at the IRCCS Hospital San Raaele. This dataset was collected in the Fleury private clinics from Sao Paulo of Brazil, containing 1301 patients' data with 352 COVID-19 positive, and 949 COVID-19 negative patients. It has 345 patient instances with 334 COVID-19 positive and 11 COVID-19 negative patients collected from the Hospital Srio-Libanes in Sao Paulo. Both Brazil-1 & 2 were collected between February 2020 and June 2020. This dataset contains 200 COVID-19 patients' data, which were collected between January and March 2021 at the National Reference Laboratory for Clinical Chemistry, Millennium COVID-19 Treatment and Care Center, the Ethiopian Public Health Institute in Addis Ababa. Table 1 shows the dataset information for COVID-19 and normal classes in detail.
Table 1

Dataset description for Model development, internal and external validation.

DatasetCOVID-19Non-COVIDTotal
aTraining (80% of OSR)Italy (OSR)6296701299
aBalanced Training Set6706701340
aInternal validation (20% of OSR)157168325
External validationItaly-1163174337
Italy-2104145249
Italy-3118106224
Brazil-13529491301
Brazil-233411345
Ethiopia200200

Five-fold cross validation was used for performance evaluation. Number of samples per fold for training, augmented training, and internal validation is reported here.

Dataset description for Model development, internal and external validation. Five-fold cross validation was used for performance evaluation. Number of samples per fold for training, augmented training, and internal validation is reported here.

Statistical analysis

We performed all statistical analysis using Python 3.7, where age, continuous variables, and other clinical data were reported (with a mean and standard deviation) for each biomarker for COVID-19 and Normal groups. The Chi-square univariate test and rank-sum test were employed to find the statistically significant features in both groups. We determined the significance with the P-value < 0.05. Gender, age, and thirteen different features were identified from the database. Table 2 summarizes 14 parameters (age, and 13 clinical biomarkers) and their statistical characteristics.
Table 2

Statistical analysis of the COVID-19 and Non -COVID groups’ characteristics using the internal training dataset.

FeaturesUnitAcronymMissing rate (%)COVID-19
Non-COVID
Overall
p-value
mean ± stdmean ± stdmean ± std
AgeyearsAge061.85 ± 16.359.3 ± 22.2460.54 ± 19.61<0.05
White blood cells109/LWBC2.47.66 ± 3.889.72 ± 5.178.7 ± 4.69<0.05
Red blood cells1012/LRBC3.64.65 ± 0.684.44 ± 0.744.54 ± 0.720.112
Hemoglobing/dlHGB2.413.54 ± 1.8912.86 ± 2.0913.2 ± 2.020.545
Hematocrit%HCT2.440.27 ± 5.338.55 ± 5.739.41 ± 5.56<0.05
Mean corpuscular volumefLMCV3.686.9 ± 6.6787.5 ± 7.4187.2 ± 7.060.288
Mean corpuscular hemoglobinpg/CellMCH3.629.23 ± 2.6329.18 ± 2.829.2 ± 2.730.655
Mean corpuscular hemoglobin concentrationg Hb/dLMCHC2.433.62 ± 1.3333.32 ± 1.3633.47 ± 1.350.881
Platelets109/LPLT13.6222.73 ± 90.78246.16 ± 97.5234.5 ± 94.8<0.05
Neutrophils count109/LNET18.95.88 ± 3.67.2 ± 5.46.47 ± 4.52<0.05
Lymphocytes count109/LLYT15.21.15 ± 0.841.64 ± 1.041.37 ± 0.96<0.05
Monocytes count109/LMOT15.20.54 ± 0.60.71 ± 0.390.61 ± 0.5<0.05
Eosinophils count109/LEOT15.20.023 ± 0.080.11 ± 0.180.064 ± 0.14<0.05
Basophils count109/LBAT15.20.005 ± 0.020.029 ± 0.0520.016 ± 0.040.075
Statistical analysis of the COVID-19 and Non -COVID groups’ characteristics using the internal training dataset. In the training data, there are 786 (48.4%) COVID-19 and 838 (51.6%) Non -COVID patients. The average age of the subjects in the COVID-19 and Normal groups were 61.85 ± 16.3 and 59.3 ± 22.24, respectively. 507 (53%) and 279 (41.2%) subjects were COVID-19 positive among males and females, respectively. The mean white blood cells (WBC) of the study population were 7.66 ± 3.88 (109/L) for the COVID-19 group and 9.72 ± 5.17 (109/L) for the Non -COVID group. For monocytes count (MOT), the mean of the study population was 0.54 ± 0.6 (109/L) for the COVID-19 group and 0.71 ± 0.39 (109/L) for the Non -COVID group. Moreover, the mean lymphocytes count (LYT) of the study population was 1.15 ± 0.84 (109/L) for the COVID-19 group whereas 1.64 ± 1.04 (109/L) for the Non -COVID group.

Data preprocessing

Before extracting features from the dataset and passing it to the training steps, we followed three preprocessing steps, namely data imputation, normalization, and balancing, to condition the dataset.

Missing data imputation

Identification and missing data imputation are the primary preprocessing steps in this study. For each patient, multiple blood biomarkers were acquired. However, some parameters are missing among the patients. The data collected during admission for the identification of the infection was used for training and validation. Data imputation techniques were used for the missing data for the key predictors instead of the simplest technique removing the instances, which can cause the loss of vital and contextual information and affect the generalized representation of the dataset [26]. In recent years, machine learning-based data imputation techniques are widely used for missing value imputation. However, this approach requires a separate model for each column containing missing values. We used multivariate imputation by chained equations (MICE), which has become a popular method of handling missing data and has outperformed other imputation techniques especially for clinical data [28]. Fig. 2 illustrates the missing values for different features in the OSR (training) dataset. While the majority of the columns are well-populated, lymphocytes, monocyte, basophils, and neutrophils have some missing values. Moreover, the spark lines at the right side of the figure denote data completeness.
Fig. 2

The number of missing data for different features in the OSR dataset (Training data). The missing data are shown as spottier, and the spark-line at right shows the shape of the dataset.

The number of missing data for different features in the OSR dataset (Training data). The missing data are shown as spottier, and the spark-line at right shows the shape of the dataset.

Normalization

The success of the machine learning models is highly dependent on the input data quality for achieving generalized performance. Data normalization involves scaling or transforming the data to make each feature contribute equally during the training process. Many studies testified performance enhancement of the machine learning models employing such normalization techniques [29]. In this study, we employed Z-score normalization using the formula:where , , , and denote the new value, original value, mean, and standard deviation of the variable values in the training samples, respectively. This method transforms the data with a mean of 0 and a standard deviation of 1. Moreover, this algorithm is sensitive to outliers.

Data balancing

Data imbalance can cause bias in the machine learning models. Therefore, dataset balancing is considered another important step to achieve model performance. The synthetic minority oversampling technique (SMOTE) is a frequently used method to resolve the data imbalance [30]. In this study, the number of Non-COVID patients for training was 670 while the number of COVID patients was 629, which is slightly unequal not entirely imbalance. However, we used SMOTE to up-sample the data of the COVID class; thus, balancing both classes of the training dataset. Fifteen features were already present in the training dataset that were evaluated carefully to determine the correlation among them. Fig. 3 (A) shows the heatmap of correlation, and it was found that few features are not highly correlated. A maximum correlation of 0.97 was found between HCT and HGB. Moreover, four features had a high correlation (r > 0.85) with other features, which were removed from the dataset. Fig. 3(B) shows the correlation heatmap after removing highly correlated features. Fig. 4 shows the marginal distribution of the data in each feature.
Fig. 3

Heatmap of correlation among different features (A) using all features, and (B) removing highly correlated features.

Fig. 4

Pair plot for the distribution of the dataset.

Heatmap of correlation among different features (A) using all features, and (B) removing highly correlated features. Pair plot for the distribution of the dataset.

Feature selection

In this study, five different feature selection techniques were used namely, chi-square test, Pearson correlation coefficient, recursive feature elimination (RFE), logistic regression, and random forest to find the best feature combination for detecting COVID-19 positive patients with high probability. Moreover, we calculated the feature importance score for each feature using five feature selection techniques, then an average of the feature importance score was used as a threshold to select the features for each technique. Finally, we have selected those features which are exceeded the threshold for all five feature selection techniques to develop the classification model. Firstly, the chi-square metric between the target and the expected values was calculated and the features with the highest chi-squared values were selected. The intuition is that if a feature is independent of the target, it is uninformative for classifying observations. The chi-square test is represented by, Where is observation in a class and is the expected observations in class. We also used Pearson's correlation method for finding the association among the features. If X, Y are two features of length, N; , are the values of X, Y, and and are the means of ’s and ’s then, Recursive feature elimination (RFE) is another technique, which is often used for feature selection, which selects features recursively considering smaller feature sets. At first, the estimator takes the whole feature set into account and the importance scores of individual features are evaluated. Afterward, the least important features are excluded from the feature set. After several recursive feature exclusion steps, the desired number of features is finally chosen [31]. Moreover, we used logistic regression and random forest for feature selection, where logistic regression is often used for binary classification. Assume that we have a data set ; , where ∼ are the vectors of input features and are binary response values [32]. If p is the predicted probability P(y = 1) and is the vector of model parameters, logistic regression can represent the event's log odds as a linear model: On the other hand, Random Forest (RF) combines several decision trees, which shows better performance compared to a single tree classifier. It makes the combination by the bootstrap aggregation method during a single decision tree construction along with random data node selection [31]. For example, a decision tree with N leaves clusters the feature space into N regions Rn, 1 ≤ n ≤ N. where N and Rn respectively denote the total region number and nth region in the feature space.

Stacking-based machine learning model

In this study, a stacking-based classifier is proposed which combines multiple best performing model on a single dataset S, which consists of feature vectors () and their ground truth labels (). Here p, q denotes the number of subjects and features, respectively. At first, N number of base-level classifiers trained using the S (xi, yi) dataset; the predictions probabilities were then used to train the meta-level classifier to produce final prediction, as illustrated in Fig. 5 .
Fig. 5

Proposed stacking model architecture.

Proposed stacking model architecture. We used five-fold cross-validation to generate a training set for the meta-level classifier . Among these folds, base-level classifiers were used on four-folds, leaving one-fold for validation. Each base-level classifier predicts a probability distribution over the possible class values. Thus, using input xi, a probability distribution is created using the predictions of the base-level classifier set:where is the set of possible class values m, and denotes the probability that example x belongs to a class as estimated (and predicted) by classifier . The metalevel attributes are thus the probabilities predicted for each possible class by each of the base-level classifiers, i.e., for and .

Development and internal validation of classification model

We investigated eight machine learning classifiers in this study, namely Random Forest [33], Support Vector Machine (SVM) [34], K-nearest neighbor (KNN) [35], XGBoost [36], Adaboost [37], Gradient boosting, linear discriminant analysis (LDA) [38], and Logistic regression [39]. We used four top-ranked features to compare different classification models, which were selected by five feature selection techniques. The best three performing classifiers were selected as base learner models ( in the stacking architecture and logistic regression was used for the meta learner model ( in the second phase of the stacking model and finally produced different performance matrices based on the prediction.

Nomogram-based scoring system and external validation

Nomograms are a frequently used graphical estimation technique that reduces the statistical models into a single probability estimation of an event. In Zlotnik's Nomolog [40], a diagnosis nomogram was constructed based on multivariate logistic regression analysis. In a binary classification scenario, logistic regression provides the probability of a class ranging from 0 to 1. In this study, a nomogram-based scoring system (QCoVSML) using logistic regression was created using the probability scores from the top base learner models. Six different hospital datasets from three different countries (Italy, Brazil, and Ethiopia) were utilized as the external validation sets to evaluate the nomogram-based stacking ML system. Moreover, to compare the predicted and actual probability of COVID-19 patients, calibration curves for internal and external validation were plotted. Moreover, we used decision curve analysis (DCA) in Stata software for finalizing the ranges of threshold of individual algorithms to identify the effectiveness of individual algorithms.

Performance metrics

To evaluate the performance of the classifiers, we used the receiver operating characteristic (ROC) curves and area under the curve (AUC) along with Precision, Sensitivity, Specificity, Accuracy, and F1-Score. Moreover, we used five-fold cross-validation, which results in an 80% and 20% split for train and test sets, respectively and according to the fold number, this process is repeated 5 times to validate the entire dataset. Since different classes had a different number of instances, we used per class weighted metrics and overall accuracy. Additionally, we used the AUC value as another evaluation metric. The mathematical representation of five evaluation metrics (weighted sensitivity or recall, specificity, precision, overall accuracy, and F1 score) are shown in Equations (6), (7), (8), (9), (10): Here, true positive, true negative, false positive, and false negative are represented as TP, TN, FP, and FN, respectively.

Results

Best feature combination for prediction of COVID-19 positive patients

As mentioned earlier, five different feature selection techniques were used to select a feature combination, where we calculated feature importance using these feature selection techniques to select the final best feature combination. We used an average of the feature importance as a threshold to select the features. All of the five feature selection techniques selected four features, such as white blood cell count (WBC), monocyte count (MOT), age, and lymphocyte count (LYT) to classify COVID-19 and Non-COVID patients while other features were not selected by all five feature selection approaches, as shown in Table 3 .
Table 3

Feature ranked according to different feature selection algorithms.

FeaturePearsonChi-2RFELogisticsRandom ForestTotal
WBCImage 1Image 1Image 1Image 1Image 15
MOTImage 1Image 1Image 1Image 1Image 15
AgeImage 1Image 1Image 1Image 1Image 15
LYTImage 1Image 1Image 1Image 1Image 15
EOTImage 1Image 1Image 1Image 14
NETImage 1Image 1Image 1Image 14
RBCImage 1Image 1Image 1Image 14
HCTImage 1Image 1Image 13
PLT1Image 1Image 1Image 13
SexImage 1Image 1Image 13
BATImage 1Image 1Image 13
Feature ranked according to different feature selection algorithms.

Development and internal validation of the stacking model

The selected four features were tested on eight different ML classifiers using five-fold cross-validation to identify which models performed well to diagnose COVID-19 positive patients. Gradient boosting classifier outperformed others with weighted precision, sensitivity, F1-score, specificity, and overall accuracy of 89.86%, 89.88%, 89.88%, 89.87%, and 89.88%, respectively. The top three algorithms (Gradient Boosting, Random Forest, and XGBoost) were used for the next step. The accuracies of the three best algorithms are 89.88%, 82.91%, and 81.43% respectively. A logistic regression model was used as a meta learner in the stacking model. The stacking model provides the best performance with weighted precision, sensitivity, F1-score, specificity, and overall accuracy, of 91.44%, 91.45%, 91.45%, 91.44%, and 91.45%, respectively. Therefore, the proposed stacking model performance outweighs other state-of-the-art machine learning classifiers. The performance of the stacking model showed about 2% improvement in terms of all the different performance matrices. Table 4 compares the overall accuracies and weighted average performance of all experimented classifiers with a 95% confidence interval to diagnose COVID-19 patients. Fig. 6 clearly shows ROC curves for the different ML classifiers and the proposed stacking model, where the proposed model produced 96% AUC and outperformed different state-of-the-art ML classifiers.
Table 4

Comparison of the average performance metrics from five-fold cross-validation for different classifiers and the stacking classifier.


Overall
Weighted with 95% CI
ClassifierAccuracyPrecisionRecallF1-scoreSpecificity
Linear Discriminant Analysis (LDA)67.88 ± 2.2767.69 ± 2.2767.88 ± 2.2767.88 ± 2.2767.77 ± 2.27
XGBoost (XGB)81.43 ± 1.8981.37 ± 1.8981.43 ± 1.8981.43 ± 1.8981.39 ± 1.89
Random Forest (RF)82.91 ± 1.8382.87 ± 1.8382.91 ± 1.8382.91 ± 1.8382.74 ± 1.84
Logistic Regression (LR)68.37 ± 2.2668.63 ± 2.2668.37 ± 2.2668.37 ± 2.2668.47 ± 2.26
Support Vector Machine (SVM)62.28 ± 2.3670.03 ± 2.2362.28 ± 2.3662.28 ± 2.3661.53 ± 2.37
AdaBoost74.66 ± 2.1274.45 ± 2.1274.66 ± 2.1274.66 ± 2.1274.22 ± 2.13
K-Nearest Neighbors (KNN)79.17 ± 1.9779.11 ± 1.9879.17 ± 1.9779.17 ± 1.9779.13 ± 1.98
Gradient Boosting (GB)89.88 ± 1.4789.86 ± 1.4789.88 ± 1.4789.88 ± 1.4789.87 ± 1.47
Stacking model (GB + RF + XGB)91.45 ± 1.3691.44 ± 1.3691.45 ± 1.3691.45 ± 1.3691.44 ± 1.36
Fig. 6

ROC curves for different ML classifiers and the stacking ML model.

Comparison of the average performance metrics from five-fold cross-validation for different classifiers and the stacking classifier. ROC curves for different ML classifiers and the stacking ML model.

Nomogram-based stacking ML model and external validation

The study also developed a nomogram based on multivariate logistic regression analysis using the probability scores of the three best models (namely Gradient Boosting (M1), Random Forest (M2), and XGBoost (M3)) which identify the NON-COVID and COVID-19 subjects reliably. The relationship between these base learner models’ probability scores and the probability of COVID-19 positive patients was evaluated by the multivariate logistic regression analysis (Table 5 ). The z-value is a widely used method for determining a data position in the distribution, which is calculated by dividing the regression coefficient by its standard error. If the z-value is too big, the corresponding true regression coefficient is not 0, and the corresponding independent variable is significant.
Table 5

The logistic regression analysis to construct the stacking ML based Nomogram.

OutcomeCoef.Bootstrap Std. Err.ZP>|z|[95% conf. Interval]
Gradient Boosting (M1)6.6853140.75871428.810.0005.1982628.172367
Random Forest (M2)1.31580.47528682.770.0060.3842555−2.247345
XGBoost (M3)0.63385730.46854631.350.176−0.2844766−1.552191
cons−3.5161280.205643−17.100.000−3.919181−3.113076
The logistic regression analysis to construct the stacking ML based Nomogram. Therefore, out of 3 probability scores, for COVID-19 subjects, using the three best models, XGBoost (M3) is not a very strong predictor but Gradient Boosting (M1) and Random Forest (M2) are strong ones. P-value can determine if an X-variable has a significant relationship to the Y variable by testing the null hypothesis that the corresponding regression coefficient is 0. The X-variables with p < 0.05, have a significant relationship to Y-variables according to the null hypothesis. This is also reflecting that the probability scores of M3 are weakly related to Y-variable. Thus, to build the nomogram prediction model, these 3-probability scores were used as predictors. As illustrated in Fig. 7 , there were 6 rows in the nomogram, ranging from 1 to 3 representing the included variables. For each variable from COVID-19 or NON-COVID group, a score was derived by drawing a vertical line downward from the value on the variable axis to the “points” axis. The points of the three variables indicated the scores (row 4) and the scores were added to derive the total score, which was displayed in row 6. Afterward, to determine the probability of patients with COVID-19, a line could be drawn from the “Total Score” axis to the “Prob” axis (row 5).
Fig. 7

Multivariate logistic regression-based Nomogram to detect COVID-19 patients. Nomogram for prediction of COVID-19 was created using Gradient boost (M1), Random Forest (M2), and XGBoost (M3).

Multivariate logistic regression-based Nomogram to detect COVID-19 patients. Nomogram for prediction of COVID-19 was created using Gradient boost (M1), Random Forest (M2), and XGBoost (M3). Alternatively, the following formula can be used to calculate the nomogram score: Probability of COVID Infection (QCovSML) = 1/(1 + exp (-Linear Prediction)) (12). The corresponding classification score (QCovSML) was calculated and listed in Fig. 8 . For the development dataset (Training set), it was observed that a QCovSML score of 4.8, which denotes 40% cutoffs of classification probability provides best COVID-19 and NON-COVID stratification. Therefore, we used 4.8 as the cutoff value to classify a subject into NON-COVID and COVID-19 groups, where greater than 40% probability refers to COVID-19 infection.
Fig. 8

Nomogram scores corresponding to the classification probabilities of COVID-19 and NON-COVID subjects.

Nomogram scores corresponding to the classification probabilities of COVID-19 and NON-COVID subjects.

Performance evaluation of the nomogram-based model

Six external datasets collected from six different hospitals from three different countries (Italy, Brazil, and Ethiopia) were used to validate the nomogram-based scoring system. According to Table 6 , the multivariate logistic regression-based nomogram model performed well for internal and external validation data. For testing data (internal validation set), it produced weighted precision, sensitivity, specificity, F1-score, and overall accuracy of 89.5%, 91.12%, 90.3%, 89.71%, and 91%, respectively. For the six external datasets, the nomogram score produced weighted precision, sensitivity, specificity, F1-score, and overall accuracy of 92.02%, 95.59%, 93.73%, 90.54%, and 93.34% respectively.
Table 6

Comparison of the average and weighted performance metrics using logistic regression-based nomogram for the external datasets.



Overall
Weighted with 95% CI
DatasetCountryAccuracyPrecisionRecallF1-scoreSpecificity
Internal validationItaly (OSR)91 ± 1.4189.5 ± 1.4991.12 ± 1.3890.3 ± 1.4489.71 ± 1.54
External validationItaly-191.69 ± 1.9388.02 ± 2.2797.13 ± 1.1792.35 ± 1.8685.89 ± 2.43
Italy-295.18 ± 1.9894.63 ± 2.0897.24 ± 1.5195.92 ± 1.8392.31 ± 2.46
Italy-392.86 ± 1.9788.79 ± 2.4297.17 ± 1.2792.79 ± 1.9888.98 ± 2.4
Brazil-192.7 ± 2.7594.25 ± 2.4695.26 ± 2.2493.25 ± 2.6587.27 ± 3.52
Brazil-295.11 ± 1.2294.42 ± 2.0694.25 ± 2.1294.33 ± 2.1495.67 ± 2.11
Ethiopiaa92.5 ± 3.6592.5 ± 3.65
Average93.34 ± 2.1992.02 ± 2.2695.59 ± 2.1893.73 ± 2.3390.54 ± 2.47

Ethiopia dataset has only COVID patients.

Comparison of the average and weighted performance metrics using logistic regression-based nomogram for the external datasets. Ethiopia dataset has only COVID patients. Fig. 9 shows that the calibration plots were near to the diagonal lines for both internal and external validation, which is the indication of a reliable model. Moreover, Fig. 10 illustrates that the net benefit of every single predictor model was positive until the threshold of 0.95, which indicates that all predictors contributed to the outcome determination. In particular, the full model showed the best performance, and where, three predictors were combined in the model.
Fig. 9

Comparison of predicted and actual probability of COVID-19 patients. (A) Internal validation representation, (B) external validation representation.

Fig. 10

Comparison of decision curves analysis of different models. The net benefit balances the probability scores for COVID-19 patients.

Comparison of predicted and actual probability of COVID-19 patients. (A) Internal validation representation, (B) external validation representation. Comparison of decision curves analysis of different models. The net benefit balances the probability scores for COVID-19 patients.

Web application for QCovSML

As an extension of this study, we have developed a web application that allows the doctors to put some demographic and clinical data (white blood cell count (WBC), monocyte count (MOT), age, and lymphocyte count (LYT)) to the web interface and our AI-based application analyzes the data to predict whether the user is COVID-19 positive or not. This application was developed using Flutter, a cross-platform app (android and iOS) development framework using the Dart programming language maintained by Google. This will in essence provide us with the maximum coverage or users, quicker development and continuous integration, seamless deployment and maintenance, easier cloud integration and increase stability. In the prototype system, the user will fill in some demographic data, as well as RT-PCR test results (if available). The RT-PCR test result will not be used for model but it can be used to enhance our model in the future with a larger dataset collected through this online framework. Next, the app will ask to input only four parameters: Age, WBC, MOT, and LYT. After collecting the data from the user, it will be transferred to our server which performs pre-processing and passes through the proposed trained model to determine whether the user is COVID-19 positive or not (Fig. 11 ). The AI backend will analyze the data and return a response to the screen. The application will display the results as well as store them in a local database based on SQLite. In brief, the application can help in screening the COVID patients quickly with few blood biomarkers and can help to reduce the load on the healthcare system.
Fig. 11

Illustration of a generic framework for the COVID-19 detection tool.

Illustration of a generic framework for the COVID-19 detection tool.

Discussion

Our study showed that CBC biomarkers can provide useful information about a probable SARS-CoV2 infection, allowing clinicians to make better judgments during hospital admission. To diagnose the COVID-19 positive patients, four impactful features: Age, WBC, MOT, and LYT were identified by the ML model. These top-ranked features were used for training and validation using different ML models, and finally, the three top-ranked models were used to train a stacking model. The stacking model produced weighted precision, sensitivity, F1-score, specificity, and overall accuracy of 91.44%, 91.45%, 91.45%, 91.44%, and 91.45%, respectively, which are comparatively better than the best performing Gradient Boosting classifier with overall accuracy, and weighted precision, sensitivity, specificity, and F1-score of 89.88%, 89.86%, 89.88%, 89.88%, and 89.87%, respectively. Moreover, a nomogram-based scoring system (QCovSML) using this stacking technique was developed, which provides 91% overall accuracy. The nomogram-based scoring system was validated with six external datasets. The overall average accuracy for the six datasets is 93.34%. The QCoVSML technique requires only easy-to-obtain biomarkers, which makes it useful for the diverse environments for both in and out-patient settings. This can be particularly useful for resource-constrained countries where RT-PCR sometimes becomes scarce, and the laboratory facility becomes unavailable as well due to reagent and expert limitations. Several studies have shown that COVID-19 patients had a less total count of leukocytes and particular of its subpopulations than healthy people and Non-COVID-19 patients with various infectious illnesses [[41], [42], [43]]. As a result, the most common CBC biomarkers for identifying COVID-19 individuals, and potentially categorizing its severity are reduced neutrophil count, lymphopenia, and eosinopenia. Liu et al. [44] recently suggested that decreasing lymphocyte counts can help for predicting the severity of COVID-19 infection early. Both lymphocytes and neutrophils are essential components of our immune system since they play a vital role in infection clearance and host defense. Lymphopenia is a common symptom among COVID-19 patients, which is caused by the reduction in the number of lymphocytes in the blood [45]. We used the percentage of lymphocytes and neutrophils in this study, and similar to the previous research [46,47], found a correlation between a lower percentage of these two concentrations with extremely ill COVID-19 patients. Furthermore, bone marrow increases neutrophil production, as cytokines induce immunosuppression and lymphocyte apoptosis, which results in a rise in NLR [48]. Limited studies have shown an investigation of ML models on routine blood exams. Formica et al. [13] proposed an ML model based on CBC, with 83% sensitivity and 82% specificity for COVID-19 detection. However, the study was built on a small sample (171 patients) gathered in a short time frame. Similarly, Avila et al. [49] developed a Bayesian model, reporting 76.7% sensitivity and specificity. In our study, we have used a large CBC biomarkers dataset to develop our proposed model and used multicenter and multicountry datasets for external validation. Firstly, WBC count, monocytes (%), age, and lymphocyte (%) parameters were selected as important biomarkers out of eleven different biomarkers for COVID-19 disease prediction using five different state-of-the-art feature selection techniques. Moreover, we developed a stacking-based ML model which produced the best performance with weighted precision, sensitivity, F1-score, specificity, and overall accuracy of 91.44%, 91.45%, 91.45%, 91.44%, and 91.45%, respectively. The stacking model provides a 2% performance gain over the other state-of-the-art machine learning classifiers. Finally, developed a nomogram-based scoring system using this stacking approach and the study validated the QCovSML scoring system with six other external CBC datasets from three different countries and the average weighted precision, sensitivity, F1-score, specificity, and overall accuracy for the six external datasets are 92.02%, 95.59%, 93.73%, 90.54%, and 93.34%, respectively. This scoring system can be used for COVID-19 detection, which can be used as an alternative method to RT-PCR for the fast and cost-effective identification of COVID-19-positive patients. Several predictive models have been published in the literature reported age, and hematologic measures, among the main variables elucidating the predictive model to predict critically ill people [50,51]. Contrary to that, our model provides a valid utilization of such biomarkers as an early predictive tool for the timely and cost-effective diagnosis of COVID-19 disease. This could provide a significantly accessible patient triaging technique, which can ease the optimum staff allocation.

Conclusion

In conclusion, the proposed tool using the stacking ML model shows promising results for identifying COVID-19 positive patients with very high sensitivity using a few CBC biomarkers (WBC, Monocytes, Age, Lymphocyte count). The performance of the model reported in this study is comparable to the current gold standard for COVID-19 diagnosis, the RT-PCR test [52]. Moreover, the proposed model has a high degree of sensitivity to predict the COVID-19 positive patients in both internal and external datasets. The model was tested on completely unseen six external clinical datasets from three different countries - making this model helpful for clinicians to effectively optimize patients’ stratification, devise efficient management strategies, and better resource utilization. Finally, this ML-based solution can be applied in both in-patients and out-patients as an alternative to the RT-PCR test for fast and cost-effective identification of COVID-19-positive patients. This is especially useful in the low-resource environment, where screening of patients can be done using the proposed method and selective RT-PCR tests can be carried out to minimize RT-PCR test and selective RT-PCR tests.

Funding

This work was supported by the : UREP28-144-3-046. The statements made herein are solely the responsibility of the authors.

Dataset and code availability

The dataset used for the development and validation of this study is available at [27] and the code and models are available at https://github.com/tawsifur/QCovSML-COVID-19-detection-system-using-CBC-biomarkers.

Ethical approval

This study used seven different hospital clinical biomarker datasets [27] from three different countries: Italy, Brazil, and Ethiopia with de-identified patients' information. For all datasets, research involving human subjects complied with all relevant national and international regulations, institutional policies and was in accordance with the tenets of the Helsinki Declaration (as revised in 2013) and was approved by the authors Institutional Review Board (70/INT/2020).

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  33 in total

1.  The hemocyte counts as a potential biomarker for predicting disease progression in COVID-19: a retrospective study.

Authors:  Yufen Zheng; Ying Zhang; Hongbo Chi; Shiyong Chen; Minfei Peng; Lifei Luo; Linping Chen; Jun Li; Bo Shen; Donglian Wang
Journal:  Clin Chem Lab Med       Date:  2020-06-25       Impact factor: 3.694

2.  False Negative Tests for SARS-CoV-2 Infection - Challenges and Implications.

Authors:  Steven Woloshin; Neeraj Patel; Aaron S Kesselheim
Journal:  N Engl J Med       Date:  2020-06-05       Impact factor: 91.245

3.  Hemogram data as a tool for decision-making in COVID-19 management: applications to resource scarcity scenarios.

Authors:  Eduardo Avila; Alessandro Kahmann; Clarice Alho; Marcio Dorn
Journal:  PeerJ       Date:  2020-06-29       Impact factor: 2.984

4.  A Tool for Early Prediction of Severe Coronavirus Disease 2019 (COVID-19): A Multicenter Study Using the Risk Nomogram in Wuhan and Guangdong, China.

Authors:  Jiao Gong; Jingyi Ou; Xueping Qiu; Yusheng Jie; Yaqiong Chen; Lianxiong Yuan; Jing Cao; Mingkai Tan; Wenxiong Xu; Fang Zheng; Yaling Shi; Bo Hu
Journal:  Clin Infect Dis       Date:  2020-07-28       Impact factor: 9.079

5.  Early triage of critically ill COVID-19 patients using deep learning.

Authors:  Wenhua Liang; Jianhua Yao; Ailan Chen; Qingquan Lv; Mark Zanin; Jun Liu; SookSan Wong; Yimin Li; Jiatao Lu; Hengrui Liang; Guoqiang Chen; Haiyan Guo; Jun Guo; Rong Zhou; Limin Ou; Niyun Zhou; Hanbo Chen; Fan Yang; Xiao Han; Wenjing Huan; Weimin Tang; Weijie Guan; Zisheng Chen; Yi Zhao; Ling Sang; Yuanda Xu; Wei Wang; Shiyue Li; Ligong Lu; Nuofu Zhang; Nanshan Zhong; Junzhou Huang; Jianxing He
Journal:  Nat Commun       Date:  2020-07-15       Impact factor: 14.919

6.  Analytical sensitivity and efficiency comparisons of SARS-CoV-2 RT-qPCR primer-probe sets.

Authors:  Chantal B F Vogels; Anderson F Brito; Anne L Wyllie; Joseph R Fauver; Isabel M Ott; Chaney C Kalinich; Mary E Petrone; Arnau Casanovas-Massana; M Catherine Muenker; Adam J Moore; Jonathan Klein; Peiwen Lu; Alice Lu-Culligan; Xiaodong Jiang; Daniel J Kim; Eriko Kudo; Tianyang Mao; Miyu Moriyama; Ji Eun Oh; Annsea Park; Julio Silva; Eric Song; Takehiro Takahashi; Manabu Taura; Maria Tokuyama; Arvind Venkataraman; Orr-El Weizman; Patrick Wong; Yexin Yang; Nagarjuna R Cheemarla; Elizabeth B White; Sarah Lapidus; Rebecca Earnest; Bertie Geng; Pavithra Vijayakumar; Camila Odio; John Fournier; Santos Bermejo; Shelli Farhadian; Charles S Dela Cruz; Akiko Iwasaki; Albert I Ko; Marie L Landry; Ellen F Foxman; Nathan D Grubaugh
Journal:  Nat Microbiol       Date:  2020-07-10       Impact factor: 30.964

7.  Laboratory Findings Associated With Severe Illness and Mortality Among Hospitalized Individuals With Coronavirus Disease 2019 in Eastern Massachusetts.

Authors:  Victor M Castro; Thomas H McCoy; Roy H Perlis
Journal:  JAMA Netw Open       Date:  2020-10-01

8.  Characteristics and predictors of death among 4035 consecutively hospitalized patients with COVID-19 in Spain.

Authors:  Juan Berenguer; Pablo Ryan; Jesús Rodríguez-Baño; Inmaculada Jarrín; Jordi Carratalà; Jerónimo Pachón; María Yllescas; José Ramón Arriba
Journal:  Clin Microbiol Infect       Date:  2020-08-04       Impact factor: 13.310

View more
  4 in total

1.  A composite ranking of risk factors for COVID-19 time-to-event data from a Turkish cohort.

Authors:  Ayse Ulgen; Sirin Cetin; Meryem Cetin; Hakan Sivgin; Wentian Li
Journal:  Comput Biol Chem       Date:  2022-04-09       Impact factor: 3.737

2.  A polygenic stacking classifier revealed the complicated platelet transcriptomic landscape of adult immune thrombocytopenia.

Authors:  Chengfeng Xu; Ruochi Zhang; Meiyu Duan; Yongming Zhou; Jizhang Bao; Hao Lu; Jie Wang; Minghui Hu; Zhaoyang Hu; Fengfeng Zhou; Wenwei Zhu
Journal:  Mol Ther Nucleic Acids       Date:  2022-04-06       Impact factor: 10.183

3.  An interpretable stacking ensemble learning framework based on multi-dimensional data for real-time prediction of drug concentration: The example of olanzapine.

Authors:  Xiuqing Zhu; Jinqing Hu; Tao Xiao; Shanqing Huang; Yuguan Wen; Dewei Shang
Journal:  Front Pharmacol       Date:  2022-09-27       Impact factor: 5.988

4.  An Ensemble Learning Model for COVID-19 Detection from Blood Test Samples.

Authors:  Olusola O Abayomi-Alli; Robertas Damaševičius; Rytis Maskeliūnas; Sanjay Misra
Journal:  Sensors (Basel)       Date:  2022-03-13       Impact factor: 3.576

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.