Literature DB >> 35866773

Early prediction of diabetes by applying data mining techniques: A retrospective cohort study.

Mohammed Zeyad Al Yousef1, Adel Fouad Yasky1, Riyad Al Shammari2,3, Mazen S Ferwana4.   

Abstract

BACKGROUND: Saudi Arabia ranks 7th globally in terms of diabetes prevalence, and its prevalence is expected to reach 45.36% by 2030. The cost of diabetes is expected to increase to 27 billion Saudi riyals in cases where undiagnosed individuals are also documented. Prevention and early detection can effectively address these challenges.
OBJECTIVE: To improve healthcare services and assist in building predictive models to estimate the probability of diabetes in patients.
METHODS: A chart review, which was a retrospective cohort study, was conducted at the National Guard Health Affairs in Riyadh, Saudi Arabia. Data were collected from 5 hospitals using National Guard Health Affairs databases. We used 38 attributes of 21431 patients between 2015 and 2019. The following phases were performed: (1) data collection, (2) data preparation, (3) data mining and model building, and (4) model evaluation and validation. Subsequently, 6 algorithms were compared with and without the synthetic minority oversampling technique.
RESULTS: The highest performance was found in the Bayesian network, which had an area under the curve of 0.75 and 0.71.
CONCLUSION: Although the results were acceptable, they could be improved. In this context, missing data owing to technical issues played a major role in affecting the performance of our model. Nevertheless, the model could be used in prevention, health monitoring programs, and as an automated mass population screening tool without the need for extra costs compared to traditional methods.
Copyright © 2022 the Author(s). Published by Wolters Kluwer Health, Inc.

Entities:  

Mesh:

Year:  2022        PMID: 35866773      PMCID: PMC9302319          DOI: 10.1097/MD.0000000000029588

Source DB:  PubMed          Journal:  Medicine (Baltimore)        ISSN: 0025-7974            Impact factor:   1.817


1. Introduction

Diabetes is a major health problem in Saudi Arabia, with the second-highest rate of diabetes in the Middle East and the seventh highest in the world, with an estimated population of 7 million living with diabetes and more than 3 million with pre-diabetes.[ The prevalence of type 2 diabetes in Saudi Arabia is 32.8%; however, it is predicted to reach 35.37% in 2020, 40.37% in 2025, and 45.36% in 2030.[ Based on data from the Institute for Health Metrics and Evaluation (IHME) in Saudi Arabia, the estimated cost of diabetes in 2014 was 17 billion riyals (US $4.5 billion), with the expectation that it will increase to 27 billion riyals (US $7.2 billion) in 2030 if undiagnosed people are documented. Moreover, if pre-diabetics were to become diabetic, the cost would increase to 43 billion riyals (USD 11.43 billion). These costs include medications, visits, and laboratory tests, which vary based on the patient’s stage and complications.[ Further, due to the high costs of treatment and the expected growth rate of diabetes, Saudi Arabia will encounter a health and financial dilemma in the near future. However, prevention and early detection can effectively address these challenges and decrease costs by preventing new cases and long-term complications.[ To this end, the current evolution of information technology and the large amount of data available that could be used by applying data mining should be used. Data mining is a concept that emerged in the 1990s as a new method for data analysis and knowledge discovery. One definition of data mining is the analysis of large observational datasets to discover unsuspected relationships and to summarize the data in novel ways that are both understandable and useful to the data owner. Another definition is the process of finding previously unidentified patterns and trends in databases and using this information to build predictive models. Data mining has evolved from its beginning to include pattern recognition, clustering, classification, machine learning, artificial intelligence, and visualization.[ Unfortunately, the real application of and information on data mining applications in healthcare are usually not found in the scientific literature but on vendor websites as success stories that do not provide the complete technical details of which data mining algorithms were used and how. Examples of success stories include those on healthcare fraud prevention and the maximization of medical revenues by the detection of underdiagnosed patients.[ Additionally, we found that data mining is being used in healthcare research. In this context, there are 2 major types of predictive analytics: supervised and unsupervised. Supervised learning uses known data or information on a specific problem and produces a predictive model, whereas unsupervised learning does not require previously known data on a certain problem to train its model, but defines clusters or groups.[ In clinical medicine, data mining can be used to extract knowledge from large complex datasets assessing disease risks, supporting clinical decisions, and predicting disease development.[ A systematic review of the application of data mining techniques in the field of diabetes research, which included 17 articles, concluded that data mining is a valuable asset for diabetes researchers because it can unearth hidden knowledge from a large amount of data. Thus, it can significantly help diabetes research and, ultimately, improve the quality of healthcare for diabetic patients.[ Multiple studies have applied data-mining techniques to health data to construct predictive models for different diseases. Harper depicted the use of multiple healthcare datasets to compare classification algorithms and predict different health problems.[ He applied discriminant analysis (DA), regression models (multiple and logistic), Classification and Regression tree-based algorithms (CART), and artificial neural networks to build the model; CART (a decision tree algorithm) achieved the best overall accuracies. Another study was conducted on 395 colorectal cancer patients to predict the 5-year survival rate using 17 variables and comparing 18 algorithms. The results showed an area under the curve of more than 0.9; in other words, the accuracy of the model’s performance was greater than 90%.[ Sayad and Halkarnikar built a model to predict heart disease in patients with a sample size of 170 records and 13 attributes. They applied a multilayer perceptron neural network as a training algorithm. Their model achieved 94% accuracy with a sensitivity of 92% and specificity of 92.5%.[ A study by Daghestani and Alshammari built a diabetes prediction model using 18 attributes from the NGHA database and found 66325 instances (diabetics 64.47%, and non-diabetics 35.53%); however, they excluded pre-diabetics. In another chronic silent disease, such as hypertension, a study in Qatar found that it is possible to use noninvasive predictors through machine learning to achieve a predictive model that can achieve the targeted screening tool for such diseases.[ In this context, calculators for predicting and estimating the risk of developing diabetes are becoming more widely used. We found many such tools, but the most frequently mentioned in the literature were the American Diabetes Association (ADA) risk calculator, Cambridge Diabetes Risk Score, and Finnish Diabetes Risk Score (FINDRISC).[ Diabetes is one of the main topics of medical research because of its longevity and high cost in the healthcare system. We aimed to build a model to predict diabetic and pre-diabetic patients using demographic information and laboratory tests without the use of diagnostic tests for diabetes, hemoglobin A1c (HbA1c), random blood sugar (RBS), and fasting blood sugar (FBS), while considering that the attributes and variables are related to diabetes and are available in the database. We used a large sample size to include as much information as possible, along with multiple measures to assess the performance of our model.

2. Materials and Method

This study was conducted at the National Guard Hospital in Riyadh, Saudi Arabia. Data sets were collected from 5 hospitals that store information in the NGHA databases in the 3 highest-populated regions: the central region (Riyadh city), the western region (Jeddah and Al Madinah cities), and the eastern region (Al Ahsa and Dammam cities). This was a chart review and retrospective cohort study. All National Guard employees, dependents, and other eligible patients with available data in the NGHA database from January 2015 to January 2019 were included. Several phases were required to achieve the study objective: data collection, data preparation, data mining, model building, and model evaluation and validation.

2.1. Data collection and attribute selection

Initially, we identified the attributes that needed to be extracted from the database by reviewing the literature on diabetes (Fig. 1). The requested attributes are listed in Table 1. Complete blood count (CBC) and basic screening tests were added because almost all patients had available test results in their records. HgA1c, FBS, and RBS were requested only to classify patients as diabetic, pre-diabetic, or non-diabetic. Subsequently, they were removed from the dataset before building the model. However, other important attributes related to diabetes were unavailable (family history of diabetes, hypertension, smoking and alcohol history, and an active or sedentary lifestyle). Initially, we retrieved 1,256,898 records for demographic data (sex, age, and region), 972,239 for vital signs (systolic blood pressure [SBP], diastolic blood pressure [DBP], body mass index [BMI]), and 2,598,103 records for laboratory tests, all of which were in separate files and sheets. Table 1 summarizes the attributes used in this study.
Figure 1.

Flow chart showing the process of preparing the data.

Table 1

Attributes used in our study.

StudyRegionRBCWBCALK_PhosSodiumHDL
Current Study (41 attributes)GenderHgbMCHCAdj_CaCO2Triglyceride
AgeMPVMgAGAPPotassiumLDL
SBPHCTPhosphorusCreatinineBUNFBS
DBPMCHUric AcidALTChlorideA1c
BMIRDWT BiliASTeGFRRBS
PlateletMCVAlbuminCaCholesterol

A1c = glycated hemoglobin, ADJ_Ca = adjusted calcium, AGAP = anion gap, ALK_Phos = alkaline phosphatase, ALT = alanine transaminase, AST = aspartate aminotransferase, BMI = body mass index, BUN = blood urea nitrogen, DBP = diastolic blood pressure, eGFR = estimated glomerular filtration rate, FBS = fasting blood sugar, HCT = hematocrit, HDL = high-density lipoprotein, Hgb = hemoglobin, LDL = low-density lipoprotein, MCH = mean corpuscular hemoglobin, MCHC = mean cell hemoglobin concentration, MCV = mean corpuscular volume, MPV = mean platelet volume, RBC = red blood cells, RBS = random blood sugar, RDW = red cell distribution width, SBP = systolic blood pressure, T Bili = total bilirubin, WBC = white blood cells.

Attributes used in our study. A1c = glycated hemoglobin, ADJ_Ca = adjusted calcium, AGAP = anion gap, ALK_Phos = alkaline phosphatase, ALT = alanine transaminase, AST = aspartate aminotransferase, BMI = body mass index, BUN = blood urea nitrogen, DBP = diastolic blood pressure, eGFR = estimated glomerular filtration rate, FBS = fasting blood sugar, HCT = hematocrit, HDL = high-density lipoprotein, Hgb = hemoglobin, LDL = low-density lipoprotein, MCH = mean corpuscular hemoglobin, MCHC = mean cell hemoglobin concentration, MCV = mean corpuscular volume, MPV = mean platelet volume, RBC = red blood cells, RBS = random blood sugar, RDW = red cell distribution width, SBP = systolic blood pressure, T Bili = total bilirubin, WBC = white blood cells. Flow chart showing the process of preparing the data.

2.2. Data preparation, cleaning, and preprocessing

Data pre-processing is an essential step that affects the prediction quality. It includes missing values, smooth noisy data, identifying or removing outliers, normalization, and transformation. Initially, we combined all data based on the medical record number (MRN) and dates when the demographic and laboratory data were entered, which resulted in the gathering of 120,758 records. Healthcare data are usually not well organized and have a lot of missing data and noise; therefore, they must be prepared first. To this end, we identified the number of unique MRNs by removing duplicates from the dataset; we had 45,365 unique MRNs. We then extracted a single record with the most complete attributes for each unique MRN from the 120,758 records. The missing data included either demographic information or laboratory tests that were not retrieved properly from the database due to technical issues or were not performed or requested by the medical team for the patient. Demographic information accounted for 15% (6 out of 41) of the attributes (region, sex, age, BMI, SBP, and DBP) in this study. Patients who had 0 of 6 demographic attributes were excluded.

2.3. Labeling the patient with the diagnosis

We used the ADA guidelines to diagnose diabetes and pre-diabetes status (Fig. 2),[ and HgA1c as the main diagnostic test to label patients with diabetes, pre-diabetic, or non-diabetic. If HgA1c was missing, we used it as a diagnostic test. If FBS was missing, we used the RBS, and if RBS was missing, we labeled the patient as “non-diabetic.” The flow charts in Figures 1 and 2 demonstrate how the data were retrieved, cleaned, and prepared for the model, and Tables 2 and 3 show the distribution of the demographic variables and statistical values. Additionally, manual inspection of the data was performed to ensure that the data were consistent and accurate. The MRNs were removed after preparing the data so that no patients could be identified. After preparing the data, we obtained a sample size of 21431 unique records that included 41 attributes. Table 4 presents the number of missing values for each attribute.
Figure 2.

Flow chart showing the process of labeling each MRN with the diagnosis. MRN = medical record number.

Table 2

Distribution and statistical values of gender based on the region.

Central regionEastern regionWestern region
GenderNN %NN%NN%NN%
Female10,05146.90%698955.69%114730.98%191536.96%
Male744634.74%504540.2%108129.20%132025.48%
None available gender393418.36%5154.1%147439.81%194537.54%
Total21,431100%12,549100%3702100%5180100%
Table 3

Distribution and statistical values of diagnosis based on gender and region.

DiabeticsPrediabeticsNone diabeticsTotal
GenderNN%NN%NN%NN%
Female613247.94%188118.71%203820.27%10,05146.89%
Male455435.60%155420.87%133817.96%744634.74%
None available gender210516.46%63816.21%59115.02%393418.36%
Total12,79159.93%407319.00%456721.31%21,431100%
Table 4

The attributes and the number of missing values from the 21431 patients.

AttributeMissingMissing %AttributeMissingMissing %AttributeMissingMissing %
Region00.00%MCV322915.07%AST554825.89%
Gender293413.69%WBC366717.11%Ca664531.01%
Age293313.69%MCHC326015.21%Sodium301114.05%
SBP1610.75%Mg615928.74%CO219278.99%
DBS1610.75%Phosphors729734.05%Potassium18608.68%
BMI8614.02%Uric Acid810737.83%BUN18498.86%
Platelet318714.87%T Bili595427.78%Chloride19379.04%
RBC413919.31%Albumin621829.01%eGFR346316.16%
Hgb316414.76%ALK_Phos573226.75%Cholesterol356116.62%
MPV317114.80%Adj_Ca816438.09%HDL289413.50%
HCT317314.81%AGAP19679.18%Triglyceride358116.71%
MCH321615.01%Creatinine238911.15%LDL580327.08%
RDW412319.24%ALT535624.99%

ADJ_Ca = Adjusted Calcium, AGAP = anion gap, ALK_Phos = Alkaline phosphatase, ALT = alanine transaminase, AST = aspartate aminotransferase, BMI = body mass index, BUN = blood urea nitrogen, DBP = diastolic blood pressure, eGFR = estimated glomerular filtration rate, HCT = hematocrit, HDL = high density lipoprotein, Hgb = hemoglobin, LDL = low-density lipoprotein, MCH = mean corpuscular hemoglobin, MCHC = mean cell hemoglobin concentration, MCV = mean corpuscular volume, MPV = Mean Platelet Volume, RBC = red blood cells, RDW = red cell distribution width, SBP = systolic blood pressure, T Bili = total bilirubin, WBC = white blood cells.

Distribution and statistical values of gender based on the region. Distribution and statistical values of diagnosis based on gender and region. The attributes and the number of missing values from the 21431 patients. ADJ_Ca = Adjusted Calcium, AGAP = anion gap, ALK_Phos = Alkaline phosphatase, ALT = alanine transaminase, AST = aspartate aminotransferase, BMI = body mass index, BUN = blood urea nitrogen, DBP = diastolic blood pressure, eGFR = estimated glomerular filtration rate, HCT = hematocrit, HDL = high density lipoprotein, Hgb = hemoglobin, LDL = low-density lipoprotein, MCH = mean corpuscular hemoglobin, MCHC = mean cell hemoglobin concentration, MCV = mean corpuscular volume, MPV = Mean Platelet Volume, RBC = red blood cells, RDW = red cell distribution width, SBP = systolic blood pressure, T Bili = total bilirubin, WBC = white blood cells. Flow chart showing the process of labeling each MRN with the diagnosis. MRN = medical record number.

2.4. Missing values

Missing values were replaced with the mean values of the associated attributes. This step aims to reduce the number of values for continuous attributes. This is achieved by splitting the range of the continuous attribute into intervals. Furthermore, discretization reduces the time needed to build the prediction model and improve the prediction results.[ The age was thus discretized into 4 groups: group 1, 0–36; group 2, 36–54; group 3, 54–72; and group 4, 72–92.

2.5. Sampling

The dataset used in this study consisted of 21,431 unique records with 12,791 who experienced diabetes; however, 4567 did not have diabetes, and 4073 had pre-diabetics. The most common metric used to evaluate machine-learning techniques is accuracy; however, this measure will not work here because of the imbalanced nature of the 3 classes. In general, there are 2 ways to address this issue: oversampling the minority class or sampling the majority class. In this study, we used both undersampling and oversampling to solve the imbalanced data problem and compared the performance of both techniques. To this end, we used the Synthetic Minority Oversampling Technique (SMOTE).[ The percentage of synthetic examples generated by SMOTE from class “None Diabetes Miletus (DM)” and “Pre-Diabetic” was 100% for both classes, meaning that the number of instances of “None DM” and “Pre-Diabetic” were 9134 and 8146, respectively.

2.6. Data mining and building the model

2.6.1. Feature selection.

Feature selection is an important aspect of building a high-performance machine learning model, and one of its main benefits is reducing the data dimensionality, which in turn reduces the number of features used in building the model, and subsequently, the time needed to build the model.[ This can be achieved by choosing the most important attributes that improve the prediction accuracy. We used an automated R-based machine learning feature selection algorithm that ranks the attributes based on their information gain.[ It evaluates the importance of an attribute by measuring the entropy gain with respect to the outcome and then ranks the attributes based on their individual evaluations. Figure 3 shows the attributes after feature selection.
Figure 3.

The selected attributes according to their information gain measures.

The selected attributes according to their information gain measures. In this study, we compared the following 6 different machine learning algorithms: K-nearest neighbors (KNN) are identified from the neighbors with K similar points in the training data that are closest to the test observation. These are then classified by estimating the conditional probability of belonging to each class and choosing the class with the highest probability.[ The random tree forest (RTF) is a classification algorithm that works by forming multiple decision trees during training and outputting the class that is the mode of the classes (classification) at testing.[ Decision trees work by learning simple decision rules extracted from data features. The deeper the tree, the more complex are the decision rules and fit of the model. However, random decision forests overcome the problem of overfitting decision trees. Support vector machine (SVM) represents the instances as a set of points of 2 types in an N-dimensional place and generates an (N − 1)-dimensional hyperplane to separate those points into 2 groups.[ SVM attempts to find a straight line that separates those points into 2 types and is situated as far as possible from all those points. The naïve Bayesian classifier (BC) is based on the Bayesian theorem and is particularly suited when the dimensionality of the inputs is high.[ despite its simplicity, naïve Bayes often outperforms more complex machine learning techniques such as SVM. Bayesian network (BN) is a simple probabilistic classifier that is considered a generalization of the naive BC that removes the dependencies between variables.[ BN is designed for modeling under uncertainty where the nodes represent variables, and arcs represent direct connections between them. The BN model allows probabilistic beliefs about variables to be updated automatically as new information becomes available. Logistic regression (LR) is a type of regression that is used to predict the outcome of the categorical dependent variable. (i.e., categorical variables have a limited number of categorical values) based on 1 or more independent variables.[ All machine learning algorithms were conducted using WEKA software (version 3.8) and R-based machine learning packages (version 3.3.1).[

2.7. Model evaluation and validation

The model was evaluated using the hold-out method,[ in which the dataset was partitioned into 2 separate datasets: one for training the machine learning model and another for testing it. For the holdout method, 2 data splits were used: training with 70% of the dataset and testing with 30% of the dataset. For all classifiers, the following evaluation metrics were calculated: precision, recall, F-score root, mean squared error, and receiver operating characteristic (ROC) curve.

3. Results

The final sample size was 21,431 patients. Finally, 15 out of 41 attributes were used to build the final model, which was chosen based on the information gained by the initial model (Fig. 3). Moreover, in our sample, 46.9% were female, 34.74% were male, and 18.36% were of unknown sex (Table 2). Table 3 shows pre-diabetic, diabetic, and non-diabetic percentages for all genders, as well as the relevant trends. However, some of the records retrieved had missing attribute values. The 3 highest missing values out of the 15 attributes used in building the model were RBC, triglyceride, and eGFR (Table 4). Table 5 presents a comparison of the performances of the different classification models with and without using the SMOTE sampling method. The results of this experiment show that BN outperforms all the other classifiers in terms of precision, recall, area under the curve, F-score, root mean square error (RMSE), and accuracy.
Table 5

Comparison of the performance of the different classification models without using the synthetic minority oversampling technique.

KNN
MeasuresRFSVMLRBCBNK = 1K = 10K = 50
Precision56%57%56%59%50%52%
Recall60%61%62%59%63%55%62%60%
AUC0.670.530.670.700.710.550.620.62
F-score54%53%56%60%52%51%
RMSE0.450.440.410.440.410.550.440.44
Accuracy53%60%62%59%63%55%62%60%
Comparison of the performance of the different classification models without using the synthetic minority oversampling technique. Table 6 shows the same comparison but with the SMOTE sampling method. The results of this experiment show better results and that BN outperforms all other classifiers in terms of precision, recall, AUC, F-score, RMSE, and accuracy.
Table 6

Comparison of the performance of the different classification models using the synthetic minority over-sampling technique.

KNN
MeasuresRFSVMLRBCBNK = 1K = 10K = 50
Precision60%53%54%66%62%51%53%49%
Recall28%61%56%59%66%53%58%60%
AUC0.640.580.660.700.750.560.600.59
F-score22%54%55%56%61%52%53%48%
RMSE0.500.440.430.440.420.560.440.43
Accuracy28%61%56%59%66%53%58%60%
Comparison of the performance of the different classification models using the synthetic minority over-sampling technique.

4. Discussion

One of the goals of the Saudi Vision 2030 strategic framework in the healthcare sector is to focus on primary care, preventive medicine, and tackling chronic diseases. However, when considering mass screening for diabetes to identify undiagnosed or at-risk individuals, the costs are always considered. In this context, a study was conducted in Brazil on a population screening program for type 2 diabetes, in which 22 million capillary glucose tests were performed in individuals aged 40 years and older. They concluded that the screening program will yield a large health benefit, but higher costs compared to no screening resulted in US$ 31,147 per quality-adjusted life-year (QALY) gained.[ Using prediction models on available data from electronic health records to aid in diagnosing and identifying people at risk could be useful for early intervention and cost-effectiveness, as data are already available, and retrieving it will cost much less than conventional mass screening. We used 5 measures to assess the performance of all the classifiers. We then compared the 6 classifiers with and without SMOTE. As shown in Tables 5 and 6, BN was the best classifier with an AUC of 0.71 and 0.75, RMSE of 0.41 and 0.42, the precision of 59% and 62%, recall of 63% and 66%, the accuracy of 63% and 66%, and F-score of 60% and 61%, respectively. Thus, the results show a fair model despite the missing data in our dataset. Most prior studies on the prediction of diabetes have focused on discussing the technical aspects of machine learning, data mining, and prediction models. However, clinicians and patients are more concerned about the clinical aspects of the models, such as why the authors chose the attributes, the relevance of attributes to diabetes, and the availability of the attributes in all electronic health records, so it can be generalized to a different population.[ In comparison, our study included attributes that are relevant to diabetes or usually available in almost every electronic medical record. We excluded any diagnostic laboratory test that could diagnose diabetes, which would aid the model in identifying patients with the disease and eventually lead to high-performance bias. This was not the case in the other studies examined. Nevertheless, according to ADA guidelines, there are cut-off points for FBS and RBS levels to diagnose patients as diabetic, pre-diabetic, or non-diabetic, and a model will be able to find patterns similar to the cut-offs, which would increase its accuracy. [ Furthermore, most studies on prediction models do not include an important class: pre-diabetics. In our model, we included all 3 groups—diabetic, pre-diabetic, and non-diabetic—which ensures that no patient at risk is left out. Moreover, the data used in most models were from a specific population, whereas our data were collected from multiple hospitals in different cities. Many studies have confirmed that prediction models are a promising method for use in healthcare because of their good performance in prediction. However, not all models use a measure that can assure the significance of their model performance, as well as no bias, under, or overfitting. Table 7 summarizes a comparison of the diabetes prediction models mentioned above.
Table 7

A comparison summary of the models.

ResultsPerformance measuresValidation and training# of attributesclass# of recordsData setAuthor
RF: Recall (90%), precision (68%)Sensitivity (recall) and PPV (precision)training/testing, percentage not mentioned18Dm, Non-DM66,325NGHA 2013-2015Daghistani T[4]
PNN: accuracy 81.49%Accuracy76% training/25% testing9Dm, Non-DM768PIMASoltani Z[18]
clustering + SVM: Accuracy 98.93, sensitivity 99.33, specificity 98.73% a and AUC of 0.97Accuracy, sensitivity, specificity, AUCCross validation9Dm, Non-DM768PIMAIlango B[19]
clustering + C4.5: accuracy 92.38 %, sensitivity (90.38), and specify (93.29).Accuracy, sensitivity, specificityCross validation9Dm, Non-DM768PIMAPatil B[20]
BN: Precision (62%), Recall (66%), AUC (0.75), F-Score (61%), RMSE (0.42), and Accuracy (66%)Sensitivity (recall), PPV (precision), AUC, F score, RMSE, accuracy70% training/ 30% testing41Dm, Pre Dm, Non-DM18,181NGHA 2015–2018Current work

BN = Bayesian Network, Dm = Diabetic, Non-DM = Non-Diabetic, PNN = probabilistic neural network, Pre Dm = Pre-diabetic, RF = Random Forest, SVM = support vector machine.

A comparison summary of the models. BN = Bayesian Network, Dm = Diabetic, Non-DM = Non-Diabetic, PNN = probabilistic neural network, Pre Dm = Pre-diabetic, RF = Random Forest, SVM = support vector machine. Comparing our results with the risk calculator tools, we found that the ADA risk calculator had a sensitivity of 79%, specificity of 67%, and positive predictive value of 10%, and was validated in a Chinese population with an AUC of 0.725.[ The Cambridge Diabetes risk score had a specificity of 72%, a sensitivity of 77%, and a likelihood ratio of 2.76. The area under the receiver-operating characteristic curve was 80% and was validated in different populations, but mostly in European countries; the results were lower with an AUC ranging from 63% to 74.5%,[ FINDRISC had a sensitivity of 81%, a specificity of 76%, and a positive predictive value of 0.05. The tool was validated in other populations, and the results showed an AUC of 0.724 and 0.75.[ All risk calculator tools were validated and showed good results in identifying at-risk patients. Some of the questions were found to be similar in all risk calculator tools, indicating their importance in identifying at-risk patients. Unfortunately, although we know the importance of these questions and how they will increase the performance of our model, we could not retrieve some of these data because they were not available in the system. On the other hand, comparing the results of identifying patients in our model had an AUC of 0.71 and 0.75, which is promising and similar to the validated risk tool measures; thus, our model could be improved if we added important missing attributes. Moreover, all risk calculator tools need to be filled manually by a physician, nurse, or person who speaks English and would require either an Internet connection to access it or have it printed. This would lead to increased costs due to time loss, papers used, availability of manpower, and internet connectivity to run these tools. This increase would affect the number of patients assessed using the risk tools. Nevertheless, it is noteworthy that our model was capable of screening 21,431 patients in a couple of minutes at no cost. Integrating an automated prediction model, such as our model, into the electronic health system would give the healthcare provider a live identification of those at risk of developing diabetes. It can be sent automatically or after the patient’s health information has been reviewed by a healthcare provider, so that proper diagnostic screening tests can be performed. This eventually leads to early identification, intervention, and management for better health outcomes and reduction of disease complications and costs.

4.1. Limitations

Because of the unavailability and missing data, as shown in Table 4, our accuracy was affected, which consequently affected the feature selection for the attributes. Nevertheless, more attributes that are clinically related to the disease can be added to increase the possibility of prediction. However, because these data were unavailable in the database, they would either have to be collected manually or the stakeholders in our hospital and the health information system department would have to be contacted to implement some changes to the system that would ensure that these data are available in the future. Furthermore, we recommend identifying technical problems with the information technology department regarding missing data, adding more attributes that are related and known to increase the risk of developing diabetes from the literature to the prediction model, collaborating with other hospitals to cover more population, increasing the sample size, excluding pediatrics, and identifying better methods for imputing missing data.

5. Conclusion

The results from our model are acceptable but can be improved. Missing and unavailable data owing to technical issues play a major role in affecting the performance of our model. Nevertheless, it could still be used in preventive and health monitoring programs and as an automated mass population screening tool to identify diabetics and individuals at high risk of developing diabetes with the available data in the database and without the need for extra costs as compared to traditional methods. Moreover, with the use of better data entry and sorting methods in our health attribution datasets, the implementation of such prediction models would be more effective in the long run, although we have achieved similar outcomes to those of similar models.

Acknowledgments

The authors would like to thank KAIRMC and KACST for their cooperation and support shown by them during the study. The authors would also like to acknowledge and value the input of Prof. Sherif Sakr as he passed away recently and could not witness the final result.
  27 in total

1.  The performance of a risk score in predicting undiagnosed hyperglycemia.

Authors:  P J Park; S J Griffin; L Sargeant; N J Wareham
Journal:  Diabetes Care       Date:  2002-06       Impact factor: 19.112

Review 2.  Data-mining technologies for diabetes: a systematic review.

Authors:  Miroslav Marinov; Abu Saleh Mohammad Mosa; Illhoi Yoo; Suzanne Austin Boren
Journal:  J Diabetes Sci Technol       Date:  2011-11-01

3.  Status of the diabetes epidemic in the Kingdom of Saudi Arabia, 2013.

Authors:  Charbel El Bcheraoui; Mohammed Basulaiman; Marwa Tuffaha; Farah Daoud; Margaret Robinson; Sara Jaber; Sarah Mikhitarian; Ziad A Memish; Mohammad Al Saeedi; Mohammad A AlMazroa; Ali H Mokdad
Journal:  Int J Public Health       Date:  2014-10-08       Impact factor: 3.380

Review 4.  Data mining in healthcare and biomedicine: a survey of the literature.

Authors:  Illhoi Yoo; Patricia Alafaireet; Miroslav Marinov; Keila Pena-Hernandez; Rajitha Gopidi; Jia-Fu Chang; Lei Hua
Journal:  J Med Syst       Date:  2011-05-03       Impact factor: 4.460

Review 5.  Prediction of survival in patients with esophageal carcinoma using artificial neural networks.

Authors:  Fumiaki Sato; Yutaka Shimada; Florin M Selaru; David Shibata; Masato Maeda; Go Watanabe; Yuriko Mori; Sanford A Stass; Masayuki Imamura; Stephen J Meltzer
Journal:  Cancer       Date:  2005-04-15       Impact factor: 6.860

Review 6.  Type 2 Diabetes Mellitus in Saudi Arabia: Major Challenges and Possible Solutions.

Authors:  Asirvatham Alwin Robert; Mohamed Abdulaziz Al Dawish; Rim Braham; Maha Ali Musallam; Ayman Abdullah Al Hayek; Nasser Hazza Al Kahtany
Journal:  Curr Diabetes Rev       Date:  2017

7.  Validation of the diabetes screening tools proposed by the American Diabetes Association in an aging Chinese population.

Authors:  Yu Cho Woo; Chi Ho Lee; Carol H Y Fong; Annette W K Tso; Bernard M Y Cheung; Karen S L Lam
Journal:  PLoS One       Date:  2017-09-14       Impact factor: 3.240

8.  Predicting hypertension using machine learning: Findings from Qatar Biobank Study.

Authors:  Latifa A AlKaabi; Lina S Ahmed; Maryam F Al Attiyah; Manar E Abdel-Rahman
Journal:  PLoS One       Date:  2020-10-16       Impact factor: 3.240

9.  A new kernel-based fuzzy level set method for automated segmentation of medical images in the presence of intensity inhomogeneity.

Authors:  Maryam Rastgarpour; Jamshid Shanbehzadeh
Journal:  Comput Math Methods Med       Date:  2014-01-29       Impact factor: 2.238

10.  Cost-effectiveness of a national population-based screening program for type 2 diabetes: the Brazil experience.

Authors:  Cristiana M Toscano; Xiaohui Zhuo; Kumiko Imai; Bruce B Duncan; Carísi A Polanczyk; Ping Zhang; Michael Engelgau; Maria Inês Schmidt
Journal:  Diabetol Metab Syndr       Date:  2015-10-31       Impact factor: 3.320

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.