| Literature DB >> 34401416 |
Sai K Devana1, Akash A Shah1, Changhee Lee2, Andrew R Roney1, Mihaela van der Schaar2,3,4, Nelson F SooHoo1.
Abstract
BACKGROUND: There remains a lack of accurate and validated outcome-prediction models in total knee arthroplasty (TKA). While machine learning (ML) is a powerful predictive tool, determining the proper algorithm to apply across diverse data sets is challenging. AutoPrognosis (AP) is a novel method that uses automated ML framework to incorporate the best performing stages of prognostic modeling into a single well-calibrated algorithm. We aimed to compare various ML methods to AP in predictive performance of complications after TKA.Entities:
Keywords: AutoPrognosis; Knee replacement; Machine learning; Predictive modeling
Year: 2021 PMID: 34401416 PMCID: PMC8349766 DOI: 10.1016/j.artd.2021.06.020
Source DB: PubMed Journal: Arthroplast Today ISSN: 2352-3441
Complications.
| Complications |
|---|
| Acute myocardial infarction: index admission or within 7 d of start of index admission |
| Pneumonia: index admission or within 7 d of start of index admission |
| Sepsis, septicemia, shock: index admission or within 7 d of start of index admission |
| Pulmonary embolism: index admission or within 30 d of start of index admission |
| Surgical site bleeding: index admission or within 30 d of index admission |
| Mechanical complications: index admission or within 90 d or start of index admission |
| Periprosthetic joint infection/wound infection: index admission or within 90 d of start of index admission |
Patient demographics and overall complications.
| Variable | All patients (n = 156,750) |
|---|---|
| Age range of patients (y) | 18-100 |
| Mean age ± SD (y) | 68.2 ± 9.2 |
| Median age (y) | 68 |
| Males | 60464 (38.6%) |
| Females | 96286 (61.4%) |
| Race | |
| Black | 8764 (5.6%) |
| Native American | 580 (0.4%) |
| Asian or Pacific Islander | 8832 (5.6%) |
| White | 116954 (74.6%) |
| Other | 18260 (11.6) |
| Unknown | 3360 (2.1%) |
| Ethnicity | |
| Hispanic | 29480 (18.8%) |
| Non-Hispanic | 125521 (80.1%) |
| Unknown | 1749 (1.1%) |
| Hospital volume range | 1-8149 |
| Mean hospital volume ± SD | 1854.6 ± 1568 |
| Insurance | |
| Medicare | 93461 (59.6%) |
| Medical | 10264 (6.5%) |
| Workers compensation | 5398 (3.4%) |
| Other | 2016 (1.3%) |
| Private | 45771 (29.2%) |
| Comorbidities (CMS Clinical Conditions) | |
| Metastatic cancer | 164 (0.1%) |
| Other major cancer | 1920 (1.2%) |
| Neoplasms | 1250 (0.8%) |
| Diabetes mellitus | 32991 (21%) |
| Malnutrition | 672 (0.4%) |
| Morbid obesity | 16818 (10.7%) |
| Rheumatoid arthritis | 6713 (4.2%) |
| Osteoarthritis | 2626 (1.7%) |
| Osteoporosis | 14226 (9.1%) |
| Dementia | 1529 (1%) |
| Major psychiatric disorder | 7537 (4.8%) |
| Paralysis | 232 (0.1%) |
| Coronary artery disease or dangina | 13668 (8.7%) |
| COPD | 7890 (5%) |
| Renal failure | 12469 (7.9%) |
| Decubitus ulcer | 89 (5.7 × 10-2%) |
| Vertebral fracture | 36 (2.3 × 10-2%) |
| Skeletal deformities | 27 (1.7 × 10-2%) |
| Posttraumatic OA | 76 (4.8 × 10-2%) |
| Total complications | 1109 (0.7%) |
| # of Patients having at least 1 complication | 989 (0.6%) |
| AMI | 91 (5.8 × 10-2%) |
| Pneumonia | 474 (0.3%) |
| Sepsis | 201 (0.1%) |
| PE | 273 (0.2%) |
| Surgical site bleeding | 13 (6.3 × 10-4%) |
| Mechanical complications | 31 (2.0 × 10-2%) |
| Infection | 26 (1.7 × 10-2%) |
AMI, acute myocardial infarction; COPD, chronic obstructive pulmonary disease; PE, pulmonary embolism; SD, standard deviations.
Hospital volume is the total number of TJA cases performed between October 01, 2015, to December 13, 2017.
Figure 1A schematic depiction of AutoPrognosis. AutoPrognosis is an automated framework that configures an optimally performing ensemble of ML-based prognostic models (various pipelines) to build a single well-calibrated algorithm for risk prediction.
List of classification methods in AutoPrognosis.
| Classification methods | ||
|---|---|---|
| Logistic regression | Random forest | Gradient boosting |
| eXtreme Gradient Boosting (XGBoost) | AdaBoost | Bagging |
| Bernoulli Naïve Bayes | Gaussian Naïve Bayes | Multinomial Naïve Bayes |
| Perceptron | Decision Trees | Support Vector Machine (SVM) |
| Latent Dirichlet Allocation (LDA) | Quadratic Discriminant Analysis (QDA) | K-Nearest Neighbors (KNN) |
| Neural Networks | ||
List of the 10 pipelines fitted to the TKA cohort.
| Pipeline # | Methods | Hyperparameters | Weight |
|---|---|---|---|
| 1 | Random Forest | (max_depth = 5, n_estimators = 98) | 0.199 |
| 2 | Random Forest | (max_depth = 5, n_estimators = 96) | 0.191 |
| 3 | Random Forest | (max_depth = 5, n_estimators = 102) | 0.170 |
| 4 | Random Forest | (max_depth = 3, n_estimators = 101) | 0.155 |
| 5 | Logistic Regression | (l2-penalty, 0.139) | 0.091 |
| 6 | Logistic Regression | (l2-penalty, 0.231) | 0.075 |
| 7 | AdaBoost | (n_estimators = 150) | 0.063 |
| 8 | XGBoost | (max_depth = 5, n_estimators = 153) | 0.045 |
| 9 | Logistic Regression | (l2-penalty, 0.029) | 0.007 |
| 10 | Gradient Boosting | (max_depth = 5, n_estimators = 96) | 0.005 |
Discriminative power and calibration.
| Methods | AUROC overall (n = 156,750) | Brier score overall (n = 156,750) | AUROC obesity (n = 16,818) | AUROC diabetes (n = 32,991) |
|---|---|---|---|---|
| Logistic Regression | 0.629 ± 0.01 (0.604-0.654) | 0.006 ± 0 (0.0063-0.0063) | 0.619 ± 0.03 (0.602-0.636) | 0.583 ± 0.07 (0.526-0.640) |
| XGBoost | 0.601 ± 0.03 (0.578-0.624) | 0.006 ± 0.0002 (0.0063-0.0066) | 0.567 ± 0.03 (0.540-0.594) | 0.590 ± 0.05 (0.549-0.630) |
| Gradient Boosting | 0.662 ± 0.04 (0.625-0.698) | 0.022 ± 0.0031 (0.0051-0.0106) | 0.634 ± 0.04 (0.601-0.666) | 0.637 ± 0.05 (0.594-0.680) |
| AdaBoost | 0.657 ± 0.03 (0.630-0.684) | 0.007 ± 0 (0.0072-0.0072) | 0.625 ± 0.02 (0.609-0.641) | 0.635 ± 0.03 (0.605-0.665) |
| Random Forest | 0.545 ± 0.02 (0.525-0.565) | 0.008 ± 0.0002 (0.0073-0.0077) | 0.534 ± 0.03 (0.508-0.559) | 0.549 ± 0.05 (0.505-0.593) |
| AutoPrognosis | 0.679 ± 0.04 (0.642-0.716) | 0.007 ± 0.0010 (0.0058-0.0075) | 0.660 ± 0.02 (0.646-0.674) | 0.657 ± 0.04 (0.620-0.693) |
All values reported as mean ± standard deviation with (95% confidence interval).
Figure 2Calibration plot. Calibration, measure of how close the predicted risk is to the observed risk, is similar between AutoPrognosis and Logistic Regression.
Figure 3Feature importance. The 15 most important binary features to AutoPrognosis are shown with their respective LR feature importance (a). Autoprognosis and logistic regression have differing feature importance for binary (a) variables as well as continuous (b) and categorical (c) variables. This suggests important nonlinear relationships that are captured by AutoPrognosis that Logistic Regression cannot. COPD, chronic obstructive pulmonary disease; DM, diabetes mellitus; CAD, coronary artery disease.
AUPRC performance (mean and 95% CI).
| Models | Aurpc |
|---|---|
| Logistic Regression | 0.015 (0.011-0.018) |
| XGBoost | 0.013 (0.010-0.015) |
| Gradient Boosting | 0.020 (0.013-0.027) |
| AdaBoost | 0.022 (0.016-0.028) |
| Random Forest | 0.008 (0.007-0.009) |
| AutoPrognosis | 0.025 (0.018-0.032) |
Confusion matrix for AutoPrognosis.
| Auto prognosis | True condition | |
|---|---|---|
| Prediction condition | True positive | False positive |
| False negative | True negative | |
Confusion matrix for logistic regression.
| Logistic regression | True condition | |
|---|---|---|
| Prediction condition | True positive | False positive |
| False negative | True negative | |