| Literature DB >> 35177700 |
Reetam Ganguli1,2, Jordan Franklin3, Xiaotian Yu4, Alice Lin5,2, Daithi S Heffernan6,7,8,9.
Abstract
Surgical management for gynecologic malignancies often involves hysterectomy, often constituting the most common gynecologic surgery worldwide. Despite maximal surgical and medical care, gynecologic malignancies have a high rate of recurrence following surgery. Current machine learning models use advanced pathology data that is often inaccessible within low-resource settings and are specific to singular cancer types. There is currently a need for machine learning models to predict non-clinically evident residual disease using only clinically available health data. Here we developed and tested multiple machine learning models to assess the risk of residual disease post-hysterectomy based on clinical and operative parameters. Data from 3656 hysterectomy patients from the NSQIP dataset over 14 years were used to develop models with a training set of 2925 patients and a validation set of 731 patients. Our models revealed the top postoperative predictors of residual disease were the initial presence of gross abdominal disease on the diaphragm, disease located on the bowel mesentery, located on the bowel serosa, and disease located within the adjacent pelvis prior to resection. There were no statistically significant differences in performances of the top three models. Extreme gradient Boosting, Random Forest, and Logistic Regression models had comparable AUC ROC (0.90) and accuracy metrics (87-88%). Using these models, physicians can identify gynecologic cancer patients post-hysterectomy that may benefit from additional treatment. For patients at high risk for disease recurrence despite adequate surgical intervention, machine learning models may lay the basis for potential prospective trials with prophylactic/adjuvant therapy for non-clinically evident residual disease, particularly in under-resourced settings.Entities:
Mesh:
Year: 2022 PMID: 35177700 PMCID: PMC8854708 DOI: 10.1038/s41598-022-06585-x
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Flow chart showing the hysterectomy patient cohort selection, model training, and performance evaluation processes. A total of 3656 patients were used for model development and were randomly divided into an 80% training set (2925 patients) and a 20% testing set (731 patients). k-fold cross-validation and grid searching for hyperparameters was conducted in the training set, and model performance was evaluated based on area under receiver operating characteristic curves and accuracy rates.
Clinical history/surgical information summary table.
| Characteristic | No residual cancer, N = 2972 | Residual cancer, N = 684 | p-value1 | OR | 95% Cl |
|---|---|---|---|---|---|
| 0.12 | |||||
| 0 | 896 (30%) | 196 (29%) | – | – | |
| 1 | 501 (17%) | 101 (15%) | 0.91 | 0.67, 1.22 | |
| 2 | 840 (28%) | 211 (31%) | 1.03 | 0.80, 1.32 | |
| 3 | 436 (15%) | 90 (13%) | 0.95 | 0.69, 1.30 | |
| 4+ | 299 (10%) | 86 (13%) | 1.38 | 0.99, 1.92 | |
| 0.33 | |||||
| No | 2176 (73%) | 514 (75%) | – | – | |
| Yes | 796 (27%) | 170 (25%) | 0.85 | 0.68, 1.06 | |
| 0.57 | |||||
| No | 1717 (58%) | 404 (59%) | – | – | |
| Yes | 1255 (42%) | 280 (41%) | 0.95 | 0.78, 1.16 | |
| < 0.001 | |||||
| No | 2755 (93%) | 666 (97%) | – | – | |
| Yes | 217 (7.3%) | 18 (2.6%) | 0.58 | 0.33, 0.97 | |
| < 0.001 | |||||
| No | 2755 (93%) | 666 (97%) | – | – | |
| Yes | 217 (7.3%) | 18 (2.6%) | |||
| 0.40 | |||||
| None | 2920 (98%) | 677 (99%) | – | – | |
| Inflammatory | 45 (1.5%) | 6 (0.9%) | 0.75 | 0.27, 1.77 | |
| Tube-ovarian abscess | 7 (0.2%) | 1 (0.1%) | 1.17 | 0.06, 7.69 | |
| Uterine weight (g) | 207 ± 13 | 186 ± 23 | 0.11 | 1.00 | 1.00, 1.00 |
OR odds ratio, CI confidence interval.
1Continuous Variable: one-way ANOVA, Binary Variable: Fisher’s Exact Test, Categorical Variable: Chi-Squared Test.
Cancer related variables summary table.
| Characteristic | No residual cancer, N = 2972 | Residual cancer, N = 684 | p-value1 | OR | 95% Cl |
|---|---|---|---|---|---|
| 0.016 | |||||
| Less than 1 cm | 315 (11%) | 97 (14%) | – | – | |
| 1–2 cm | 442 (15%) | 87 (13%) | 0.59 | 0.41, 0.85 | |
| Greeter than 2 cm | 2215 (75%) | 500 (73%) | 0.65 | 0.49, 0.87 | |
| 0.12 | |||||
| No | 2256 (76%) | 499 (73%) | – | – | |
| Yes | 716 (24%) | 185 (27%) | 1.07 | 0.86, 1.34 | |
| < 0.001 | |||||
| No | 2207 (74%) | 292 (43%) | – | – | |
| Yes | 765 (26%) | 392 (57%) | 1.67 | 1.36, 2.04 | |
| < 0.001 | |||||
| No | 1651 (56%) | 129 (19%) | – | – | |
| Yes | 1321 (44%) | 555 (81%) | 1.90 | 1.49, 2.43 | |
| < 0.001 | |||||
| No | 2817 (95%) | 562 (82%) | – | – | |
| Yes | 155 (5.2%) | 122 (18%) | 1.56 | 1.15, 2.10 | |
| < 0.001 | |||||
| No | 2886 (97%) | 619 (90%) | – | – | |
| Yes | 86 (12.9%) | 65 (9.5%) | 1.04 | 0.71, 1.52 | |
| < 0.001 | |||||
| No | 2651 (89%) | 360 (53%) | – | – | |
| Yes | 321 (11%) | 324 (47%) | 3.57 | 2.88, 4.44 | |
| < 0.001 | |||||
| No | 1410 (47%) | 152 (22%) | – | – | |
| Yes | 1562 (53%) | 532 (78%) | 1.81 | 1.46, 2.26 | |
| < 0.001 | |||||
| 0–1B2 | 164 (5.5%) | 3 (0.4%) | – | – | |
| II–IVB | 100 (3.4%) | 16 (2.3%) | 3.07 | 0.91, 14.1 | |
| Not a cervical cancer case | 2708 (91%) | 665 (97%) | 3.86 | 1.37, 16.2 | |
| < 0.001 | |||||
| 0–II | 561 (19%) | 19 (2.8%) | – | – | |
| IIA–IIIC | 580 (20%) | 142 (21%) | 3.88 | 2.33, 6.81 | |
| IV–IVB | 29 (1.0%) | 14 (2.0%) | 6.19 | 2.60, 14.5 | |
| Not a corpus uteri cancer case | 1802 (61%) | 509 (74%) | 2.93 | 1.74, 5.18 | |
| < 0.001 | |||||
| I–III | 560 (19%) | 42 (6.1%) | – | – | |
| IIIA–IV | 1050 (35%) | 457 (67%) | 2.23 | 1.56, 3.24 | |
| Not an ovarian cancer case | 1362 (46%) | 185 (27%) | 1.68 | 1.09, 2.62 | |
OR odds ratio, CI confidence interval.
1Continvous Variable: one-way ANOVA, Binary Variable: Fisher’s Exact Test, Categorical Variable: Chi-Squared Test.
Clinical complication outcome variables summary table.
| Characteristic | No residual cancer, N = 2972 | Residual cancer, N = 684 | p-value1 | OR | 95% CI |
|---|---|---|---|---|---|
| 0.007 | |||||
| No | 2874 (97%) | 646 (94%) | – | – | |
| Yes | 98 (3.3%) | 38 (5.6%) | 0.83 | 0.50, 1.36 | |
| < 0.001 | |||||
| No | 2769 (93%) | 576 (84%) | – | – | |
| Yes | 203 (6.8%) | 108 (16%) | 1.79 | 1.29, 2.49 | |
| 0.68 | |||||
| No | 2941 (99%) | 675 (99%) | – | – | |
| Yes | 31 (1.0%) | 9 (1.3%) | 0.47 | 0.18, 1.13 | |
| 0.96 | |||||
| No | 2962 (100%) | 681 (100%) | – | – | |
| Yes | 10 (0.3%) | 3 (0.4%) | 0.61 | 0.12, 2.45 | |
| > 0.99 | |||||
| No | 2970 (100%) | 683 (100%) | – | – | |
| Yes | 2 (< 0.1%) | 1 (0.1%) | 3.21 | 0.12, 52.7 | |
| 0.52 | |||||
| No | 2969 (100%) | 682 (100%) | – | – | |
| Yes | 3 (0.1%) | 2 (0.3%) | 3.30 | 0.40, 23.4 | |
OR odds ratio, CI confidence interval.
1Continuous Variable: one-way ANOVA, Binary Variable: Fisher’s Exact Test, Categorical Variable: Chi-Squared Test.
Figure 2Analysis of the importance of each variable in the XGBoost machine learning model. The histogram describes the relative importance of all 35 clinical features in the logistic regression model. The relative importance is quantified by assigning a weight between 0 and 100 for each variable.
Comparative chart displaying the accuracy score, area under the receiver operating characteristic curve, F1 score, and Matthews Correlation Coefficient (MCC) for each individual machine learning model.
| Model | Accuracy | ROC AUC | F1 | MCC |
|---|---|---|---|---|
| Logistic regression | 88% (85–90%) p < 0.05 | 0.89 (0.86–0.92) p < 0.05 | 0.74 (0.69–0.78) p < 0.05 | 0.50 (0.42–0.58) p < 0.05 |
| Random forest | 88% (86–90%) p < 0.05 | 0.88 (0.84–0.91) p < 0.05 | 0.72 (0.67–0.77) p < 0.05 | 0.52 (0.44–0.60) p < 0.05 |
| XGBoost | 87% (85–90%) p < 0.05 | 0.88 (0.85–0.91) p < 0.05 | 0.73 (0.69–0.78) p < 0.05 | 0.51 (0.42–0.59) p < 0.05 |
| K-nearest neighbors | 81% (80–86%) p < 0.05 | 0.72 (0.67–0.76) p < 0.05 | 0.53 (0.49–0.58) p < 0.05 | 0.19 (0.09–0.28) p < 0.05 |
| Support vector machine | 81% (80–85%) p < 0.05 | 0.50 (0.44–0.56) p < 0.05 | 0.45 (0.44–0.46) p < 0.05 | 0.0 (0.00–0.00) p < 0.05 |
A 95% confidence interval is listed for each metric in parentheses, and a p-value expressing if a metric for each model is significantly different from the same metric from all the other models is listed.
Figure 3Evaluation of the machine learning models’ predictive abilities: receiver operating characteristic curves of all models plotted, along with area under the curve for each model listed in the legend.