| Literature DB >> 30367589 |
Hsin-Yi Tsao1,2, Pei-Ying Chan3,4, Emily Chia-Yu Su5,6.
Abstract
BACKGROUND: The risk factors of diabetic retinopathy (DR) were investigated extensively in the past studies, but it remains unknown which risk factors were more associated with the DR than others. If we can detect the DR related risk factors more accurately, we can then exercise early prevention strategies for diabetic retinopathy in the most high-risk population. The purpose of this study is to build a prediction model for the DR in type 2 diabetes mellitus using data mining techniques including the support vector machines, decision trees, artificial neural networks, and logistic regressions.Entities:
Keywords: Clinical decision support; Diabetic retinopathy; Machine learning; Risk factors
Mesh:
Year: 2018 PMID: 30367589 PMCID: PMC6101083 DOI: 10.1186/s12859-018-2277-0
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Statistical analysis of categorical variables
| Value | Retinopathy | |||||
|---|---|---|---|---|---|---|
| DM | Normal | |||||
| Count | Percentage (%) | Count | Percentage (%) | |||
| Exercise | Y | 60 | 56.60 | 39 | 36.79 | 0.3266 |
| N | 46 | 43.40 | 67 | 63.21 | ||
| Family history | Y | 60 | 56.60 | 37 | 34.91 | 0.2054 |
| N | 46 | 43.40 | 69 | 65.09 | ||
| Insulin | Y | 40 | 37.74 | 10 | 9.43 | < 0.0001* |
| N | 66 | 62.26 | 96 | 90.57 | ||
| SMBG | Y | 67 | 63.21 | 61 | 57.55 | 0.3995 |
| N | 39 | 36.79 | 45 | 42.45 | ||
| Gender | F | 51 | 48.11 | 62 | 58.49 | 0.1300 |
| M | 55 | 51.89 | 44 | 41.51 | ||
Counts and percentages of categorical variables between DR and normal patients are calculated
*Variables with p-value < 0.05 are highlighted
Statistical analysis of numerical variables
| Retinopathy | |||||||||
|---|---|---|---|---|---|---|---|---|---|
| DM | Normal | ||||||||
| Min | Max | Mean | StdDev | Min | Max | Mean | StdDev | ||
| SBP | 96 | 223 | 137.89 | 18.76 | 101 | 196 | 132.25 | 15.72 | 0.0188* |
| DBP | 55 | 112 | 78.75 | 10.39 | 28 | 101 | 75.73 | 11.24 | 0.0435* |
| BMI | 17.6 | 38.1 | 25.99 | 3.75 | 19.6 | 49.3 | 27.35 | 5.08 | 0.0278* |
| Age | 35 | 88 | 61.50 | 10.77 | 19 | 84 | 57.36 | 12.92 | 0.0120* |
| ≧65 | 65 | 88 | 71.63 | 5.25 | 65 | 84 | 72.25 | 4.89 | 0.6054* |
| 40–64 | 43 | 64 | 55.51 | 5.82 | 41 | 64 | 53.88 | 5.88 | 0.1217* |
| < 40 | 35 | 39 | 36.75 | 2.06 | 19 | 38 | 28.75 | 6.31 | 0.0363* |
| Duration | 1 | 36 | 12.88 | 7.93 | 1 | 23 | 7.50 | 5.18 | < 0.001* |
Minimum (Min), maximum (Max), mean, and standard deviation (StdDev) of numerical variables between DR and normal patients are calculated
*Variables with p-value < 0.05 are highlighted
Prediction performance using percentage split
| Model | Training | Test | ||||||
|---|---|---|---|---|---|---|---|---|
| AUC | Acc. | Sens. | Spec. | AUC | Acc. | Sens. | Spec. | |
| SVM | 0.783 | 0.708 | 0.787 | 0.664 |
|
|
| 0.724 |
| LR | 0.749 | 0.679 | 0.703 | 0.660 | 0.802 | 0.727 | 0.813 | 0.679 |
| ANN | 0.875 | 0.762 | 0.756 | 0.768 | 0.777 | 0.682 | 0.682 | 0.682 |
| DT | 0.719 | 0.685 | 0.660 | 0.718 | 0.768 | 0.727 | 0.708 |
|
AUC, accuracy, sensitivity, and specificity of different machine learning algorithms using training (i.e., 80%) and test (i.e., 20%) data sets are evaluated
aBest evaluation measures in test set are underlined
Fig. 1ROC plots for the training and test data sets. ROC curves of different machine learning algorithms (i.e., DT, LR, SVM, and ANN) for the training (80%) and test (20%) data sets
Prediction performance using five-fold cross-validation
| Model | Five-fold cross-validation | |||
|---|---|---|---|---|
| AUC | Acc. | Sens. | Spec. | |
| SVM | 0.821 | 0.791 | 0.819 | 0.782 |
| LR | 0.756 | 0.763 | 0.761 | 0.742 |
| ANN | 0.738 | 0.731 | 0.692 | 0.727 |
| DT | 0.690 | 0.718 | 0.683 | 0.729 |
AUC, Accuracy, sensitivity, and specificity of different machine learning algorithms using five-fold cross-validation are evaluated
Prediction performance using three-way data split
| Training | Validation | Test | ||||
|---|---|---|---|---|---|---|
| Model | Acc. | AUC | Acc. | AUC | Acc. | AUC |
| SVM | 0.863 | 0.961 |
|
|
|
|
| LR | 0.831 | 0.769 | 0.813 | 0.707 | 0.798 | 0.712 |
| ANN | 0.872 | 0.849 | 0.794 | 0.707 | 0.780 | 0.685 |
| DT | 0.825 | 0.707 | 0.817 | 0.693 | 0.780 | 0.640 |
Accuracy and AUC of different machine learning algorithms using training (i.e., 60%), validation (i.e., 20%), and test (i.e., 20%) data sets are evaluated
aBest evaluation measures in validation set are underlined as selected mode and independent performance evaluation is shown in bold
Fig. 2ROC plots for the training, validation, and test data sets. ROC curves of different machine learning algorithms (i.e., DT, LR, SVM, and ANN) for the training (60%), validation (20%), and test (20%) data sets
Performance of previous studies
| Approaches | Data Sets | AUC | Acc. | Sens. | Spec. |
|---|---|---|---|---|---|
| Hosseini et al. | Iran | 0.704 | NAa | 0.603 | 0.694 |
| Oh et al. | South Korea | 0.820 | 0.752 | 0.721 | 0.760 |
| Ogunyemi et al. | United States | 0.720 | 0.735 | 0.692 | 0.559 |
AUC, accuracy, sensitivity, and specificity of the best predictive performance reported in previous studies are summarized
aNA stands for “Not Available” because this evaluation measure was not reported in the study
Performance comparison with previous studies
| Approaches | Data Sets | Patients | Features | AUC | Acc. | Sens. | Spec. |
|---|---|---|---|---|---|---|---|
|
| |||||||
| Hosseini et al. | Taiwan | 212 | 10 |
|
|
| 0.689 |
| Iran | 3734 | 11 | 0.704 | NAb | 0.603 |
| |
|
| |||||||
| Oh et al. | Taiwan | 212 | 10 |
|
|
| 0.757 |
| South Korea | 490 | 37 | 0.820 | 0.752 | 0.721 |
| |
|
| |||||||
| Ogunyemi et al. | Taiwan | 212 | 10 |
| 0.667 | 0.682 |
|
| United States | 513 | 24 | 0.720 |
|
| 0.559 | |
AUC, accuracy, sensitivity, and specificity of our Taiwan data set are compared with the Iran data set in Comparison 1 (i.e., using Hosseini et al.’s approach), with the South Korea data set in Comparison 2 (i.e., using Oh et al.’s approach), with the United States data set in Comparison 3 (i.e., using Ogunyemi et al.’s approach)
aBest evaluation measures in each comparison are underlined
bNA stands for “Not Available” because this evaluation measure was not reported in the study
Fig. 3Clinical intepretation using decision trees. Interpretable rules for clinical practice generated by decision tress
Odds ratio estimates of important risk factors
| Effects | Point Estimates | |
|---|---|---|
| Duration | 1.093 | |
| Insulin | Y vs. N | 3.561 |
Odds ratio estimates of duration and insulin variables generated by logistic regression model
Performance comparison of different years in duration
| Model | Training Acc. | Test Acc. |
|---|---|---|
| DT (10-yr) | 0.649 |
|
| DT (15-yr) | 0.601 | 0.659 |
| DT (2-yr) | 0.512 | 0.500 |
Ranked prediction performance of decision trees based on 2-year, 10-year, and 15-year duration of diabetes