| Literature DB >> 29668729 |
Sherif Sakr1,2,3, Radwa Elshawi4,3, Amjad Ahmed1,2, Waqas T Qureshi5, Clinton Brawner6, Steven Keteyian6, Michael J Blaha7, Mouaz H Al-Mallah1,2,6.
Abstract
This study evaluates and compares the performance of different machine learning techniques on predicting the individuals at risk of developing hypertension, and who are likely to benefit most from interventions, using the cardiorespiratory fitness data. The dataset of this study contains information of 23,095 patients who underwent clinician- referred exercise treadmill stress testing at Henry Ford Health Systems between 1991 and 2009 and had a complete 10-year follow-up. The variables of the dataset include information on vital signs, diagnosis and clinical laboratory measurements. Six machine learning techniques were investigated: LogitBoost (LB), Bayesian Network classifier (BN), Locally Weighted Naive Bayes (LWB), Artificial Neural Network (ANN), Support Vector Machine (SVM) and Random Tree Forest (RTF). Using different validation methods, the RTF model has shown the best performance (AUC = 0.93) and outperformed all other machine learning techniques examined in this study. The results have also shown that it is critical to carefully explore and evaluate the performance of the machine learning models using various model evaluation methods as the prediction accuracy can significantly differ.Entities:
Mesh:
Year: 2018 PMID: 29668729 PMCID: PMC5905952 DOI: 10.1371/journal.pone.0195344
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Dataset description.
| 49 +/- 12 | |
| 12,694 (55%) | |
| 10,401 (45%) | |
| 4694 (20%) | |
| 18401 (80%) | |
| 12581 (54%) | |
| 1956 (8%) | |
| 255 (1%) | |
| 524 (2%) | |
| 2286 (10%) | |
| 1004 (4%) | |
| 10.2 +/- 2.79 | |
| 124 +/- 17 | |
| 79 +/- 10 | |
| 73 +/- 12 | |
| 82 +/- 13 | |
| 159 +/- 17 | |
| 1887 (8%) | |
| 9,518 (41%) | |
| 11,865 (51%) | |
| 7,769 (34%) | |
| 314 (1%) | |
Fig 1A flowchart of our experimental process.
Fig 2The information gain ranking of the attributes of the dataset.
Fig 3AUC of different models with different percentage of synthetic examples created using SMOTE evaluated using 10-fold cross validation.
Fig 4AUC of the different ML models using Spread Subsample technique.
Comparison of the performance of Support Vector Machine (SVM) classifier with sampling using polynomial, normalized polynomial and puk kernels using complexity parameters 0.1, 10 and 30 using 10-fold cross validation using SMOTE.
| Polynomial | Normalized Polynomial | Puk | |||||||
|---|---|---|---|---|---|---|---|---|---|
| C = 0.1 | C = 10 | C = 30 | C = 0.1 | C = 10 | C = 30 | C = 0.1 | C = 10 | C = 30 | |
| 46.55 | 46.70 | 46.66 | 40.86 | 45.05 | 44.62 | 48.61 | 58.74 | 63.61 | |
| 77.39 | 77.31 | 77.33 | 79.17 | 77.40 | 78.09 | 78.44 | 78.90 | 78.97 | |
| 52.57 | 52.56 | 52.57 | 51.36 | 51.76 | 52.30 | 54.83 | 59.97 | 61.95 | |
| 49.38 | 49.45 | 49.44 | 45.51 | 48.18 | 48.15 | 51.53 | 59.35 | 62.77 | |
| 0.62 | 0.62 | 0.62 | 0.60 | 0.61 | 0.61 | 0.64 | 0.69 | 0.71 | |
| 0.58 | 0.58 | 0.58 | 0.59 | 0.58 | 0.58 | 0.57 | 0.53 | 0.51 | |
Comparison of the performance of Artificial Neural Networks (ANN) classifier with gradient descent back-propagation using hidden units {1, 2, 4, 8} and the momentum {0,0.2, 0.5} using 10-fold cross validation using SMOTE.
| H = 1 | H = 2 | H = 4 | H = 8 | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| M = 0 | M = 0.2 | M = 0.5 | M = 0 | M = 0.2 | M = 0.5 | M = 0 | M = 0.2 | M = 0.5 | M = 0 | M = 0.2 | M = 0.5 | |
| 59.30 | 50.06 | 51.44 | 49.77 | 51.33 | 42.21 | 30.06 | 32.29 | 62.05 | 51.38 | 56.41 | 37.49 | |
| 64.34 | 73.22 | 71.68 | 72.45 | 71.52 | 79.07 | 88.00 | 85.04 | 57.74 | 73.50 | 66.83 | 82.34 | |
| 47.24 | 50.16 | 49.44 | 49.30 | 49.25 | 52.05 | 57.43 | 53.75 | 44.15 | 51.07 | 47.79 | 53.33 | |
| 52.59 | 50.11 | 50.42 | 49.53 | 50.27 | 46.61 | 39.46 | 40.34 | 51.59 | 51.22 | 51.74 | 44.03 | |
| 0.66 | 0.66 | 0.65 | 0.66 | 0.66 | 0.67 | 0.67 | 0.67 | 0.66 | 0.67 | 0.67 | 0.67 | |
| 0.47 | 0.47 | 0.47 | 0.47 | 0.47 | 0.46 | 0.46 | 0.46 | 0.47 | 0.47 | 0.47 | 0.47 | |
Comparison of the performance of Bayesian Network classifier (BN) using different search algorithms K2, Hill Climbing, Repeated Hill Climber, LAGD Hill Climbing, TAN, Tabu and Simulated Annealing using 10-fold cross validation using SMOTE.
| K2 | Hill Climbing | Repeated Hill Climber | LAGD Hill Climbing | TAN | Tabu | Simulated Annealing | |
|---|---|---|---|---|---|---|---|
| 45.29 | 45.24 | 45.24 | 33.49 | 40.42 | 44.95 | 36.26 | |
| 79.91 | 79.90 | 79.90 | 85.01 | 83.36 | 79.94 | 85.86 | |
| 54.83 | 54.79 | 54.79 | 54.60 | 56.67 | 54.68 | 57.99 | |
| 49.60 | 49.56 | 49.56 | 41.51 | 47.19 | 49.34 | 44.62 | |
| 0.70 | 0.70 | 0.70 | 0.67 | 0.70 | 0.70 | 0.70 | |
| 0.47 | 0.47 | 0.47 | 0.46 | 0.45 | 0.47 | 0.45 |
Fig 5AUC Curves for the Different Machine Learning Models using SMOTE evaluated using 10-fold cross-validation.
Fig 6AUC Curves for the Different Machine Learning Models using SMOTE and evaluated using holdout (70/30).
Fig 7AUC Curves for the Different Machine Learning Models using SMOTE and evaluated using holdout (80/20).
The performance of the Different Machine Learning Models evaluated using the 10-fold cross validation method using SMOTE.
The RTF model achieves the highest AUC (0.93), F-Score (86.70%), sensitivity (69,96%) and Specificity (91.71%).
| ANN | LB | LWB | RTF | BN | SVM | |
|---|---|---|---|---|---|---|
| 30.06% | 31.28% | 37.22% | 69.96% | 36.26% | 63.61% | |
| 88.00% | 88.56% | 84.05% | 91.71% | 85.86% | 78.97% | |
| 57.43% | 59.53% | 55.67% | 81.69% | 57.99% | 61.95% | |
| 39.46% | 41.01% | 53% | 86.70% | 44.62% | 62.77% | |
| 0.67 | 0.69 | 0.67 | 0.93 | 0.70 | 0.71 | |
| 0.46 | 0.54 | 0.46 | 0.34 | 0.45 | 0.51 |
The performance of the Different Machine Learning Models evaluated using the Hold Out method (70/30) using SMOTE.
The RTF model achieve the highest AUC (0.88), Sensitivity (74.30%), Precision (73.50%) and F-Score (73.90%).
| ANN | LB | LWB | RTF | BN | SVM | |
|---|---|---|---|---|---|---|
| 39.50% | 31.40% | 40.80% | 74.30% | 48.80% | 26.30% | |
| 86.50% | 88.60% | 81.80% | 85.60% | 79.30% | 88.60% | |
| 61.20% | 59.80% | 54.60% | 73.50% | 55.90% | 55.50% | |
| 48% | 41.20% | 46.64% | 73.90% | 52.10% | 35.70% | |
| 0.72 | 0.70 | 0.70 | 0.88 | 0.71 | 0.58 | |
| 0.54 | 0.451 | 0.46 | 0.36 | 0.47 | 0.58 |
The performance of the Different Machine Learning Models evaluated using the Hold Out method (80/20) using SMOTE.
The RTF model achieves the highest AUC (0.89), Sensitivity (75%), Precision (73%) and F-Score (74%). The SVM model achieves the highest Specificity (88.9%).
| ANN | LB | LWB | RTF | BN | SVM | |
|---|---|---|---|---|---|---|
| 40% | 31.3% | 43% | 75% | 49.5% | 28.2% | |
| 88.4% | 88.5% | 80.92% | 86.2% | 79.8% | 88.9% | |
| 65.2% | 59.3% | 54.8% | 73% | 56.8% | 57.7% | |
| 49.8% | 40.9% | 48.23% | 74% | 52.9% | 37.9% | |
| 0.74 | 0.7 | 0.7 | 0.89 | 0.72 | 0.59 | |
| 0.44 | 0.45 | 0.46 | 0.46 | 0.42 | 0.57 |