| Literature DB >> 35941641 |
Huan Wang1, Qin-Yu Zhao2, Jing-Chao Luo1, Kai Liu1, Shen-Ji Yu1, Jie-Fei Ma1,3, Ming-Hao Luo4, Guang-Wei Hao1, Ying Su1, Yi-Jie Zhang1, Guo-Wei Tu5, Zhe Luo6,7,8.
Abstract
BACKGROUND: Noninvasive ventilation (NIV) has been widely used in critically ill patients after extubation. However, NIV failure is associated with poor outcomes. This study aimed to determine early predictors of NIV failure and to construct an accurate machine-learning model to identify patients at risks of NIV failure after extubation in intensive care units (ICUs).Entities:
Keywords: Categorical Boosting; Hyperparameter optimization; Non-invasive mechanical ventilation failure; Prospective validation; Recursive feature elimination
Mesh:
Substances:
Year: 2022 PMID: 35941641 PMCID: PMC9358918 DOI: 10.1186/s12890-022-02096-7
Source DB: PubMed Journal: BMC Pulm Med ISSN: 1471-2466 Impact factor: 3.320
Fig. 1Flow chart of patient selection. eICU-CRD, eICU Collaborative Research Database; ICU, intensive care unit
Fig. 2Schematic illustration of the study design. A Patients with NIV initiated within 48 h after extubation in the eICU Collaborative Research Database were included in the study, and 93 variables were extracted. The dataset was divided into a training set (70%) and internal validation set (30%). B The recursive feature elimination algorithms were performed on the training set, and key features were selected. C Hyperparameters was optimized by using an automated machine learning toolkit on the training set. D The developed CatBoost model outperformed other models in both the internal validation and prospective validation sets
Baseline characteristics of the eICU-CRD and PREDICt cohorts
| eICU-CRD cohort | PREDICt cohort | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Missing | Overall | NIV success | NIV failure | Missing | Overall | NIV success | NIV failure | |||
| n | 929 | 681 | 248 | 419 | 333 | 86 | ||||
Ext. To NIV ≤ 24 h, n (%) h | 0 | 876 (94.3) | 647 (95.0) | 229 (92.3) | 0.164 | 0 | 348 (83.1) | 289 (86.8) | 59 (68.6) | < 0.001 |
Age, mean (SD) years | 0 | 62 (15) | 62 (15) | 61 (15) | 0.374 | 1 | 61 (14) | 61 (14) | 63 (12) | 0.120 |
| GCS, median [Q1, Q3] | 263 | 14 [12, 15] | 15 [13, 15] | 11 [9, 15] | < 0.001 | 0 | 15 [13, 15] | 15 [13, 15] | 14 [13, 15] | < 0.001 |
Heart Rate, mean (SD) beats per min | 3 | 88 (15) | 87 (15) | 91 (16) | 0.001 | 31 | 95 (20) | 95 (20) | 97 (18) | 0.433 |
Respiratory rate, mean (SD) beats per min | 14 | 21 (5) | 20 (5) | 22 (6) | < 0.001 | 31 | 21 (7) | 21 (7) | 22 (6) | 0.370 |
MBP, mean (SD) beats per min | 15 | 82 (14) | 82 (14) | 82 (13) | 0.689 | 37 | 81 (12) | 81 (12) | 82 (11) | 0.409 |
| SpO2, median [Q1, Q3] | 64 | 97 [95, 98] | 97 [95, 98] | 97 [95, 98] | 0.312 | 47 | 99 [96, 100] | 99 [97, 100] | 98 [96, 100] | 0.048 |
Temperature, mean (SD) ℃ | 39 | 37.2 (0.6) | 37.2 (0.6) | 37.2 (0.7) | 0.574 | 8 | 37.3 (0.7) | 37.3 (0.7) | 37.2 (0.7) | 0.177 |
Glucose, mean (SD) mg/dl | 132 | 131 (39) | 129 (35) | 138 (49) | 0.010 | 59 | 160 (39) | 159 (38) | 165 (42) | 0.271 |
| pH, mean (SD) | 567 | 7.38 (0.06) | 7.38 (0.06) | 7.36 (0.08) | 0.058 | 51 | 7.44 (0.09) | 7.44 (0.08) | 7.42 (0.13) | 0.263 |
PaO2, median [Q1, Q3] mmHg | 568 | 89 [74, 113] | 91 [75, 113] | 84 [67, 106] | 0.039 | 52 | 103 [81, 137] | 99 [76, 132] | 79 [62, 107] | < 0.001 |
Urine Output, median [Q1, Q3] mL/kg/h | 3 | 0.6 [0.0, 1.2] | 0.6 [0.0, 1.2] | 0.7 [0.2, 1.4] | 0.039 | 54 | 1.8 [1.4, 2.3] | 1.8 [1.4, 2.4] | 1.8 [1.5, 2.2] | 0.823 |
Input, median [Q1, Q3] mL/kg/h | 3 | 0.1 [0.0, 0.6] | 0.0 [0.0, 0.6] | 0.2 [0.0, 0.9] | < 0.001 | 62 | 1.3 [1.1, 1.7] | 1.3 [1.1, 1.8] | 1.4 [1.1, 1.7] | 0.521 |
Mechanical Ventilation Duration, median [Q1, Q3] h | 2 | 34.5 [12.4, 96.0] | 22.7 [8.1, 51.2] | 121.0 [55.4, 227.6] | < 0.001 | 10 | 28.0 [18.0, 62.0] | 25.0 [18.0, 47.0] | 61.5 [18.2, 94.5] | < 0.001 |
Mean Airway Pressure, mean (SD) cmH2O | 555 | 10.2 (7.3) | 8.5 (6.9) | 12.9 (7.0) | < 0.001 | 0 | 9.8 (1.6) | 9.5 (1.1) | 11.1 (2.2) | < 0.001 |
| Failure Type | ||||||||||
| Tracheotomy | – | – | – | 45 (18.1) | – | – | – | – | 51 (59.3) | – |
| Reintubation | – | – | – | 210 (84.7) | – | – | – | – | 75 (87.2) | – |
| Mortality | – | – | – | 80 (32.4) | – | – | – | – | 31 (36.0) | – |
Values are presented as the mean (SD) if not otherwise specified
eICU-CRD, eICU Collaborative Research Database; ZS, Zhongshan; NIV, noninvasive ventilation; GCS, Glasgow Coma Scale Score; MBP, mean blood pressure; SpO2, saturation of peripheral oxygen; pH, potential of hydrogen; PaO2, arterial partial pressure of oxygen; NIV, noninvasive ventilation
Fig. 3Hyperparameter optimization process with an automated machine learning toolkit. A The blue point represents the result of a trail, and the dark orange line represents the best area under the receiver operating characteristic curve (AUROC). B Each line represents a trail, and the green to red color line represents its AUROC
Fig. 4Comparison of the full and compact CatBoost models. The full model was developed on the basis of all available features, whereas the compact model was derived on the basis of key features selected by the recursive feature elimination algorithm. Both models had optimized hyperparameters. A Receiver operating characteristic curves (ROCs) of the full and the compact models. Distribution of the effects of each feature on the output of B the full model or C the compact model, estimated using the SHapley Additive exPlanations (SHAP) values. The plot sorts features by the sum of SHAP value magnitudes over all samples. The blue to red color represents the feature value (red high, blue low). The x-axis measures the effects on model output (right positive, left negative)
Fig. 5Comparison of model performance with other predictive tools and in the internal validation set. A Receiver operating characteristic curves (ROCs) of CatBoost and other predictive tools/factors. B Receiver operating characteristic curves (ROCs) of different models
Model performance in the internal and prospective validation sets
| Model | AUC | ACC (%) | Best cutoff | Gray zone | Values in gray zone | Youden index (%) | Sensitivity (%) | Specificity (%) | F1 Score | PPV (%) | NPV (%) |
|---|---|---|---|---|---|---|---|---|---|---|---|
| CatBoost | 0.87 (0.82–0.92) | 76 (71–81) | 0.045 | 0.04–0.08 | 64 (22.94%) | 64 | 89 (75–100) | 75 (70–80) | 0.41 (0.29–0.52) | 27 (18–36) | 98 (96–100) |
| Random Forest | 0.85 (0.78–0.92) | 73 (68–78) | 0.077 | 0.08–0.14 | 51 (18.28%) | 64 | 92 (80–100) | 72 (66–77) | 0.39 (0.27–0.50) | 25 (16–34) | 99 (97–100) |
| XGBoost | 0.83 (0.76–0.89) | 71 (66–76) | 0.007 | 0.00–0.04 | 241 (86.38%) | 57 | 89 (76–100) | 69 (64–74) | 0.36 (0.25–0.47) | 23 (14–31) | 98 (96–100) |
| KNN | 0.82 (0.75–0.89) | 75 (70–80) | 0.05 | 0.04–0.09 | 85 (30.47%) | 62 | 89 (74–100) | 74 (68–79) | 0.40 (0.28–0.50) | 25 (17–35) | 98 (96–100) |
| GBDT | 0.82 (0.73–0.90) | 76 (71–81) | 0.03 | 0.01–0.06 | 161 (57.71%) | 52 | 77 (60–92) | 76 (70–81) | 0.37 (0.25–0.48) | 24 (15–34) | 97 (94–99) |
| NaiveBayes | 0.81 (0.72–0.88) | 66 (61–71) | 0.013 | 0.01–0.25 | 97 (34.77%) | 55 | 93 (80–100) | 63 (57–69) | 0.33 (0.24–0.43) | 20 (14–28) | 99 (97–100) |
| LightGBM | 0.80 (0.71–0.87) | 74 (69–79) | 0.004 | 0.00–0.03 | 242 (86.74%) | 54 | 81 (64–95) | 74 (68–79) | 0.37 (0.25–0.48) | 24 (15–33) | 97 (95–99) |
| LR | 0.80 (0.70–0.89) | 73 (67–78) | 0.055 | 0.02–0.12 | 155 (55.56%) | 56 | 85 (69–97) | 72 (66–77) | 0.36 (0.25–0.47) | 23 (15–32) | 98 (96–99) |
| SVM | 0.79 (0.72–0.86) | 63 (57–68) | 0.066 | 0.07–0.12 | 80 (28.67%) | 52 | 93 (80–100) | 60 (54–66) | 0.32 (0.22–0.41) | 19 (12–26) | 99 (97–100) |
| AdaBoost | 0.77 (0.66–0.86) | 85 (81–89) | 0.486 | 0.47–0.49 | 118 (42.29%) | 45 | 58 (38–76) | 88 (83–92) | 0.42 (0.27–0.55) | 33 (20–47) | 95 (92–98) |
| COX | 0.75 (0.64–0.84) | 71 (66–76) | 0.242 | 0.15–0.43 | 135 (48.39%) | 47 | 77 (59–93) | 70 (65–76) | 0.32 (0.22–0.43) | 21 (13–29) | 97 (94–99) |
| MLP | 0.73 (0.62–0.83) | 66 (60–71) | 0.001 | 0.00–0.04 | 241 (86.38%) | 38 | 74 (55–91) | 65 (59–71) | 0.28 (0.18–0.38) | 17 (11–25) | 96 (93–99) |
| CatBoost | 0.85 (0.80–0.89) | 72 (68–77) | 0.062 | 0.06–0.11 | 85 (20.29%) | 62 | 92 (84–98) | 70 (65–74) | 0.44 (0.35–0.52) | 29 (22–36) | 98 (97–100) |
| XGBoost | 0.81 (0.76–0.86) | 67 (63–72) | 0.014 | 0.01–0.15 | 149 (35.56%) | 53 | 88 (78–96) | 64 (60–69) | 0.40 (0.31–0.47) | 26 (19–32) | 98 (95–99) |
| GBDT | 0.81 (0.76–0.85) | 66 (61–70) | 0.047 | 0.04–0.18 | 150 (33.94%) | 51 | 88 (78–96) | 63 (58–68) | 0.36 (0.28–0.44) | 23 (17–29) | 98 (96–99) |
| Random Forest | 0.80 (0.75–0.85) | 74 (70–78) | 0.167 | 0.11–0.23 | 164 (37.10%) | 50 | 76 (64–88) | 74 (70–78) | 0.40 (0.31–0.49) | 27 (20–35) | 96 (94–98) |
| COX | 0.76 (0.70–0.81) | 67 (62–71) | 0.38 | 0.37–0.60 | 151 (34.16%) | 52 | 88 (79–96) | 64 (59–69) | 0.37 (0.30–0.45) | 24 (18–30) | 98 (96–99) |
| LightGBM | 0.74 (0.67–0.80) | 68 (63–72) | 0.013 | 0.00–0.09 | 358 (85.44%) | 34 | 67 (54–80) | 68 (63–72) | 0.33 (0.25–0.41) | 22 (16–29) | 94 (91–97) |
| AdaBoost | 0.72 (0.64–0.79) | 67 (63–72) | 0.483 | 0.47–0.49 | 275 (62.22%) | 41 | 74 (61–86) | 66 (62–71) | 0.34 (0.26–0.42) | 22 (16–29) | 95 (92–98) |
| KNN | 0.70 (0.63–0.77) | 68 (64–72) | 0.039 | 0.01–0.07 | 272 (61.54%) | 31 | 62 (49–75) | 69 (64–73) | 0.30 (0.22–0.38) | 20 (14–27) | 93 (90–96) |
| LR | 0.68 (0.60–0.76) | 79 (75–83) | 0.085 | 0.02–0.11 | 273 (61.76%) | 34 | 52 (37–65) | 82 (78–86) | 0.36 (0.25–0.45) | 27 (18–36) | 93 (90–96) |
| NaiveBayes | 0.67 (0.59–0.74) | 65 (61–70) | 0.021 | 0.00–0.15 | 387 (87.56%) | 28 | 62 (50–76) | 66 (61–70) | 0.29 (0.21–0.36) | 19 (14–25) | 93 (90–96) |
| MLP | 0.66 (0.58–0.74) | 58 (54–63) | 0.005 | 0.00–0.27 | 391 (88.46%) | 32 | 76 (64–88) | 56 (51–61) | 0.29 (0.22–0.36) | 18 (13–23) | 95 (92–98) |
| SVM | 0.65 (0.58–0.72) | 51 (46–55) | 0.055 | 0.04–0.14 | 305 (69.00%) | 32 | 86 (75–95) | 46 (41–51) | 0.28 (0.21–0.35) | 17 (12–22) | 96 (93–99) |
Models are ordered according to the area under the receiver operating characteristic curve. The Youden index was defined as sensitivity + specificity − 1.
XGBOOST, eXtremely Gradient Boosting; GBDT, Gradient Boosting Decision Tree; KNN, K-Nearest Neighbor; SVM, Support Vector Machine; MLP, Multi-Layer Perceptron; LR, Logistic Regression; PPV, positive predictive value; NPV, negative predictive value
Fig. 6Application of the CatBoost model. A Receiver operating characteristic curves of different models in the prospective validation set. B Influence of the SHAP value on model output. C An example of the web-based tool. D The prediction results of CatBoost model and the top ten importance features are summarized. A green bar indicates a protective factor, whereas a red bar represents a risk factor. Bar length corresponds to the magnitude of protection or risk