Literature DB >> 30282956

Derivation and Validation of Machine Learning Approaches to Predict Acute Kidney Injury after Cardiac Surgery.

Hyung-Chul Lee1, Hyun-Kyu Yoon2, Karam Nam3, Youn Joung Cho4, Tae Kyong Kim5, Won Ho Kim6, Jae-Hyon Bahk7.   

Abstract

Machine learning approaches were introduced for better or comparable predictive ability than statistical analysis to predict postoperative outcomes. We sought to compare the performance of machine learning approaches with that of logistic regression analysis to predict acute kidney injury after cardiac surgery. We retrospectively reviewed 2010 patients who underwent open heart surgery and thoracic aortic surgery. Baseline medical condition, intraoperative anesthesia, and surgery-related data were obtained. The primary outcome was postoperative acute kidney injury (AKI) defined according to the Kidney Disease Improving Global Outcomes criteria. The following machine learning techniques were used: decision tree, random forest, extreme gradient boosting, support vector machine, neural network classifier, and deep learning. The performance of these techniques was compared with that of logistic regression analysis regarding the area under the receiver-operating characteristic curve (AUC). During the first postoperative week, AKI occurred in 770 patients (38.3%). The best performance regarding AUC was achieved by the gradient boosting machine to predict the AKI of all stages (0.78, 95% confidence interval (CI) 0.75⁻0.80) or stage 2 or 3 AKI. The AUC of logistic regression analysis was 0.69 (95% CI 0.66⁻0.72). Decision tree, random forest, and support vector machine showed similar performance to logistic regression. In our comprehensive comparison of machine learning approaches with logistic regression analysis, gradient boosting technique showed the best performance with the highest AUC and lower error rate. We developed an Internet⁻based risk estimator which could be used for real-time processing of patient data to estimate the risk of AKI at the end of surgery.

Entities:  

Keywords:  acute kidney injury; cardiovascular surgery; machine learning

Year:  2018        PMID: 30282956      PMCID: PMC6210196          DOI: 10.3390/jcm7100322

Source DB:  PubMed          Journal:  J Clin Med        ISSN: 2077-0383            Impact factor:   4.241


1. Introduction

Generalized linear models, such as logistic regression analysis, have been used to predict postoperative morbidity. However, the logistic regression model requires the statistical assumption of a linear relationship between the covariates and the risk of morbidity. Furthermore, the limitation of overfitting and multicollinearity of regression analysis preclude the analysis of many explanatory variables. These limitations have restricted the analysis model to select a small set of variables that are known to be clinically relevant. Recently, the machine learning technique has been applied in areas of medicine, including detecting a specific clinical finding on medical imaging and has shown excellent performance with high sensitivity and specificity [1,2]. Additionally, there were reports about the use of machine learning techniques to predict postoperative clinical outcomes including specific morbidity or in-hospital mortality [3,4,5]. Machine learning techniques showed better performance and low error rates to predict clinical outcomes compared to the logistic regression or Cox regression analysis. However, there was also a study reporting that the machine learning technique did not show a better performance that a previous risk prediction model for in-hospital mortality [5]. Postoperative acute kidney injury (AKI) after cardiovascular surgery is known to be a relevant complication because it is associated with increased long-term mortality and development of chronic kidney disease [6,7,8]. To find a risk factor and develop a risk prediction model, previous studies reported the results of multivariable logistic regression analysis [9,10,11,12,13,14,15,16,17]. Although many risk factors and risk scores were reported by multivariable logistic regression analysis, their performance in terms of the area under the receiver operating characteristic curves (AUC) was about 0.70 to 0.83 with room for further improvement [9,10,13,14,18]. Furthermore, previous prediction models may have included an insufficient number of perioperative variables owing to overfitting and multi-collinearity of the logistic regression analysis. Additionally, the potential non-linear relationship between the covariates and the risk of outcome cannot be considered. However, machine learning techniques are relatively free of these limitations of statistical analysis and may demonstrate better performance than that of logistic regression analysis. Therefore, we attempted to directly compare the performance and error rate of prediction with machine learning techniques with that of prediction with multivariable logistic regression analysis. We hypothesized that prediction with machine learning techniques involving many perioperative variables may demonstrate better performance and low error rate than that of logistic regression analysis. We evaluated as many machine learning techniques as possible that are currently available in the statistical software package R (version 3.4.4., R Development Core Team, Vienna, Austria) because the R software package is easily and freely accessible to investigators and many packages for machine learning approaches are currently available.

2. Materials and Methods

2.1. Study Design

This retrospective observational study was approved by the institutional review board of Seoul National University Hospital (1805-170-948). We retrospectively reviewed the electronic medical records of 2010 consecutive patients who underwent coronary artery surgery, valve replacement, or thoracic aortic surgery at our institution between 2008 and 2015. The need for informed consent was waived because of the retrospective design of the study.

2.2. Anesthesia, Surgical Technique

General anesthesia was maintained using a target-controlled infusion of propofol and remifentanil, or inhalational anesthetics during the study period. Standard monitoring devices were applied, including pulmonary artery catheters (Swan-Ganz CCOmbo CCO/SvO2™; Edward Lifesciences LLC, Irvine, CA, USA), in all patients.

2.3. Data Collection

On the basis of previous studies, data related to demographic or perioperative variables known to be related to postoperative renal dysfunction were collected (Table 1) [6,9,10,11,12,13,14,15,16,17,19,20,21,22,23]. The following perioperative clinical variables were collected: patient demographics, medical history, medication history, baseline laboratory finding, surgery type, operation time, type of anesthesia, intraoperative fluid and colloid administration, intraoperative transfusion amount, and intraoperative hemodynamic variables.
Table 1

Patient characteristics and postoperative renal function in the dataset.

VariablesAllTraining SetTest Setp-Value
Patient population, n201010051005
Demographic data
Age (years)64 (56–71)64 (56–71)64 (55–71)0.884
Female (n)553 (27.5)279 (27.8)274 (27.3)0.803
Body-mass index (kg/m2)23.8 (21.6–25.9)23.9 (21.7–25.9)23.7 (21.5–25.9)0.563
Surgery type
Coronary artery bypass (n)911 (45.3)473 (47.1)438 (43.6)0.117
Valvular heart surgery (n)1052 (52.3)503 (50.0)549 (54.6)0.060
Thoracic aortic surgery (n)47 (2.3)29 (2.9)18 (1.8)0.104
Emergency (n)51 (2.5)26 (2.6)25 (2.5)0.887
Previous cardiac surgery (n)149 (7.4)75 (7.5)74 (7.4)0.932
Medical history
Hypertension (n)1057 (52.6)538 (53.5)519 (51.6)0.396
Diabetes mellitus (n)588 (29.3)302 (30.0)286 (28.5)0.433
Three vessel disease (n)602 (30.0)306 (30.4)296 (29.5)0.626
Previous coronary stent insertion (n)235 (11.7)118 (11.7)117 (11.6)0.945
Cerebrovascular accident (n)228 (11.3)101 (10.0)127 (12.6)0.078
COPD (n)100 (5.0)49 (4.9)51 (5.1)0.837
Pulmonary hypertension (n)129 (6.4)60 (6.0)69 (6.9)0.413
Chronic kidney disease (n)121 (6.0)57 (5.7)64 (6.4)0.512
Preoperative Medication
ACEi (n)114 (5.7)58 (5.8)56 (5.6)0.847
ARB (n)249 (12.4)122 (12.1)127 (12.6)0.735
β-blocker (n)289 (19.4)199 (19.8)190 (18.9)0.611
Diuretics (n)297 (14.8)133 (13.2)164 (16.3)0.059
Calcium channel blocker (n)287 (14.3)151 (15.0)136 (13.5)0.339
Statins (n)506 (25.2)255 (25.4)251 (25.0)0.837
Aspirin (n)957 (47.6)498 (49.6)459 (45.7)0.090
Baseline laboratory findings
Preoperative LVEF (%)58 (52–63)58 (53–63)57 (52–63)0.427
Hematocrit (%)38 (34–42)38 (34–42)38 (34–42)0.844
Serum creatinine (mg/dL)0.94 (0.80–1.12)0.93 (0.80–1.10)0.94 (0.80–1.13)0.613
Serum Albumin (g/dL)4.1 (3.8–4.3)4.1 (3.9–4.3)4.1 (3.8–4.3)0.183
Serum uric acid (mg/dL)4.6 (3.7–5.6)4.6 (3.7–5.7)4.5 (3.6–5.5)0.190
Blood glucose (mg/dL)115 (96–146)116 (96–146)113 (96–147)0.500
Surgery and anaesthesia details
Operation time (h)6.25 (5.33–7.25)6.25 (5.41–7.27)6.25 (5.33–7.24)0.654
Anesthesia time (h)7.50 (6.25–8.50)7.50 (6.50–8.50)7.50 (6.50–8.42)0.608
Total intravenous anesthesia (n)1858 (92.4)937 (93.2)921 (91.6)0.206
Inhalational anesthesia (n)152 (7.6)68 (6.8)84 (8.4)0.206
Intraoperative crystalloid infusion (L)2150 (1150–3000)2200 (1100–3100)2150 (1200–2950)0.656
Intraoperative colloid use (mL)900 (350–1500)1000 (350–1550)800 (350–1500)0.067
pRBC transfusion during surgery (units)2 (0–3)2 (0–3)2 (0–3)0.725
FFP transfusion during surgery (units)0 (0–3)0 (0–3)0 (0–3)0.589
Intraoperative mean arterial pressure (mmHg)72 (67–78)72 (67–78)72 (67–78)0.974
Intraoperative mean cardiac index (L/min)2.3 (2.1–2.7)2.3 (2.1–2.7)2.3 (2.1–2.7)0.257
Intraoperative mean SvO2 (%)73 (69–76)73 (69–76)73 (68–76)0.207
Intraoperative diuretics use (n)204 (10.1)91 (9.1)113 (11.2)0.107
Postoperative renal function
AKI according to KDIGO criteria (n) 0.596
Stage 1591 (29.4)282 (28.1)309 (30.7)
Stage 2114 (5.7)60 (6.0)54 (5.4)
Stage 365 (3.2)33 (3.3)32 (3.2)
Hemodialysis dependent (n)125 (6.2)60 (6.0)65 (6.5)0.644
GFR at postoperative day one (ml/min/1.73m2)79 (58–94)79 (57–95)78 (58–94)0.864

Data are presented as median (interquartile range) or number (%). COPD = chronic obstructive pulmonary disease, ACEi = angiotensin-converting-enzyme inhibitor, AKI = acute kidney injury, ARB = angiotensin II receptor blocker, LVEF = left ventricular ejection fraction, pRBC = packed red blood cell transfusion, FFP = fresh-frozen plasma, SvO2 = mixed venous oxygen saturation, KDIGO = kidney disease improving global outcomes, GFR = glomerular filtration rate.

The primary outcome variable was postoperative AKI defined according to the Kidney Disease Improving Global Outcomes (KDIGO) criteria, which was determined according to the maximal change in serum creatinine level during the first seven postoperative days [6,24]. The most recent serum creatinine level measured before surgery was used as the baseline value. The detailed diagnostic criteria are shown in Table S1. We did not use the urine output criteria because previous studies suggested that different cutoffs of oliguria may be required for AKI after surgery [25,26]. We also analyzed the stage 2 or 3 AKI as secondary outcomes because stage 1 AKI may only be transient and functional and stage 2 or 3 AKI is more strongly associated with patient mortality [27]. The prediction of severe stages of AKI would be practically more important.

2.4. Statistical Analysis

R software version 3.4.4. (R Development Core Team, Vienna, Austria) was used for our analysis. The following R packages for machine learning approaches were used: Tree, rpart, ROSE (Random Over-Sampling Examples), randomForest, DMwR (Data Mining with R), XGBoost (eXtreme Gradient Boosting), e1071, UBL (utility-based learning), Kernlab, nnet, neuralnet, and h2o. Tree, rpart, and ROSE packages with CART (Classification And Regression Tree) analysis were used for decision tree analysis; randomForest and DMwR were used for random forest; XGboost was used for extreme gradient boosting; e1071, UBL, and kernlab were used for support vector machine; nnet and neuralnet were used for neural network regression; and h2o was used for deep belief networks (Text S1). Seventy-two explanatory variables including variables in Table 1 were used to machine learning. Our sample was randomly divided into a training and test set with a ratio of 1:1. The coefficients of machine learning techniques were trained with the training set and tested with the test set. Our primary analysis attempted to compare the predictive accuracy of machine learning approaches with traditional analytic techniques for classification, and previous risk scores for AKI after cardiac surgery [9,10,11,12,13,14,15,16]. To evaluate and compare the predictive accuracy of prediction by machine learning techniques and logistic regression models, we calculated the areas under the receiver operating characteristics curve (AUCs) [28,29] and compared AUCs of all classifiers and models using De Long’s method [30]. We also compared the error rate, which was defined as the sum of the number of cases with false positive and false negative divided by the size of the test set. The error rates of the logistic regression model and other previous risk scores were calculated by using a cutoff where the sum of sensitivity and specificity was maximal. For decision tree analysis, the number of terminal nodes was determined considering the scree plot showing the relationship between the tree size and coefficient of variance. We considered several decision trees with some terminal nodes that were associated with a small coefficient of variance. The final decision tree model that is clinically acceptable was chosen. The decision tree was pruned based on cross-validated error results using the complexity parameter associated with the minimal error. The ROSE package generates a synthetic balanced dataset with both over- and under-sampling and allows strengthening of the subsequent estimation of any binary outcomes [31]. The randomForest package provided a variable importance plot which shows the relative importance of the explanatory variables according to the mean decrease in accuracy or Gini. DMwR package is a technique to improve predictive ability by increasing the number of positive cases, which is called SMOTE (Synthetic Minority Over-sampling Technique). The XGBoost provides extreme and efficient gradient boosting [32,33,34]. The e1071 package was used for the support vector machine. The UBL package provides an over-sampling technique of SMOTE, which was also used to handle the class imbalance in the training set for the support vector machine [35]. The parameters of the support vector machine for classification was tuned based on balance data after SMOTE. The best parameters were determined to be a gamma of 0.1 at a cost of 10. The kernlab package provided the least square support vector machine. The neuralnet package provided the neural network classification and the number of hidden layers was defined as 6 with minimal error. The h2o deep learning package was used for deep learning. Multivariable logistic regression analysis, including the variables in Table 1, was performed to identify independent predictors used for the development of a multivariable prediction model. To avoid multicollinearity, variables that were closely correlated with each other were excluded before being entered into the multivariable analysis. Backward stepwise variable selection was conducted using cutoff of p < 0.10. Previous risk scores of Palomba, Wijeysundera, Mehta, Thakar, Brown, Aronson, Fortescue, and Rhamanian et al. [9,10,11,12,13,14,15,16] were also applied to our study data and their performance was also compared with logistic models of ours, as well as other machine learning techniques. As a sensitivity analysis, logistic regression analysis without stepwise variable selection was performed to evaluate the performance. Missing data were noted in <8% of records. We imputed the missing values according to the incidence of the missing values for each predictor. If the incidence of the missing was <2%, the missing values were substituted by the mean of continuous variables and by the mode for the incidence variable. The missing values of variables with a missing ratio of >2% and <8% were replaced using multiple imputations. Multiple imputations were performed separately in the training and test dataset. Multiple imputed training and test datasets were combined for a single run of the machine learning classifiers or logistic regression analysis. We developed a risk estimator based on our gradient boosting model [36]. This estimator calculates the risk of developing AKI after cardiac surgery and classifies the risk into three classes of low, moderate, and high risk of AKI.

3. Results

A total of 2010 cases including 911 (45.3%) coronary artery bypass and 1052 (52.3%) valve replacement surgery cases were included in our analysis. During the first seven postoperative days, AKI, as determined according to the KDIGO criteria, was observed in 770 patients (38.3%) and stage 2 or 3 AKI developed in 179 patients (8.9%). The incidence of AKI was 37.3% (375/1005) for the training set and 39.3% (395/1005) for the test set. The incidences of stage 2 or 3 AKI were 9.3% (93/1005) and 8.6% (86/1005) for training and test set, respectively. Patient demographics and surgery-related variables in both training and test set are compared in Table 1. The error rate and AUCs of all machine techniques, logistic regression model, and risk scores to predict AKI of all stages in the test data set were compared in Table 2 and Figure 1. Extreme gradient boosting classification showed the lowest test error rate (26.0%) and the largest AUC (0.78, 95% confidence interval (CI) 0.75–0.80), which was significantly greater than AUCs of other machine learning techniques or risk scores compared (p < 0.001). The deep belief network classifier showed the highest test error rate (47.2%) and smallest test AUC (0.55) among all machine learning techniques compared. The error rate and AUCs to predict AKI of stage 2 or 3 in the test set were compared in Table S2. Gradient boosting classification showed lowest test error rate (8.5%) and the largest AUC (0.74). The results of multivariable logistic regression analysis with and without stepwise variable selection was shown in Table 3 and Table S3. The AUC of the multivariable logistic prediction model with stepwise variable selection was 0.69 (95% CI 0.66 to 0.72) and the model without variable selection showed similar AUC (Table 2).
Table 2

Comparison of area under receiver-operating characteristic curve among the different models.

ModelSoftware or R PackagesError Rate of Test Data SetAUC in the Test Set
Machine learning techniques
Decision tree, CARTtree, rpart28.9%0.71 (0.67–0.74)
ROSE decision treeROSE30.6%0.66 (0.65–0.72)
Random forest modelrandomForest30.4%0.68 (0.64–0.71)
Random forest SMOTE modelDMwR33.5%0.68 (0.65–0.71)
Gradient boosting classificationXGBoost26.0%0.78 (0.75–0.80) *
Support vector machine, classifiere107131.4%0.67 (0.63–0.70)
Support vector machine, SMOTE modelUBL33.3%0.68 (0.65–0.71)
Support vector machine, least squareKernlab30.2%0.69 (0.66–0.72)
Neural network classifiernnet38.4%0.64 (0.61–0.68)
Neural network classifierneuralnet43.9%0.57 (0.53–0.61)
Deep belief networkh2o47.2%0.55 (0.51–0.59)
Risk scores from logistic regression analysis
Logistic regression model, stepwise variable selectionR33.6%0.69 (0.66–0.72)
Logistic regression model, without variable selectionR32.8%0.70 (0.68–0.73)
AKICS scoreR43.4%0.57 (0.53–0.60)
Wijeysundera and colleaguesR45.2%0.55 (0.51–0.59)
Metha and colleaguesR45.8%0.55 (0.52–0.59)
Thakar and colleaguesR45.3%0.56 (0.53–0.60)
Brown and colleaguesR43.1%0.58 (0.54–0.61)
Aronson and colleaguesR43.3%0.58 (0.51–0.62)
Fortescue and colleaguesR44.2%0.56 (0.52–0.60)
Rhamanian and colleaguesR47.0%0.55 (0.52–0.58)

Error rate was defined as sum of the number of cases with false positive and false negative divided by all test set. * Significantly greater than AUC of all the other techniques, AUC = area under the receiver operating characteristic curve, CART = Classification And Regression Tree, ROSE = Random Over-Sampling Examples, SMOTE = Synthetic Minority Over-sampling Technique, DMwR = Data Mining with R, XGBoost = eXtreme Gradient Boosting, UBL = utility-based learning, AKICS = acute kidney injury following cardiac surgery.

Figure 1

Comparison of AUC among the different machine learning models and logistic regression model. AKICS = acute kidney injury after cardiac surgery.

Table 3

Development of multivariable logistic regression model to predict acute kidney injury using stepwise variable selection.

VariableBeta-CoefficientOdds Ratio95% CIp-Value
Age per 10 year0.1281.141.04–1.610.004
History of hypertension0.3201.381.12–1.690.002
Baseline chronic kidney disease0.9072.481.62–3.78<0.001
Preoperative E/e´ > 150.4541.581.27–1.96<0.001
Preoperative hematocrit, %−0.0620.940.92–0.96<0.001
Surgery time, per 1 h0.0731.081.01–1.150.036
Intraoperative red blood cell transfusion, unit0.0561.061.01–1.110.022
Intraoperative fresh frozen plasma transfusion, unit0.0851.091.03–1.150.001
Intraoperative diuretics use0.6301.881.36–2.60<0.001

Multivariable logistic regression analysis was performed using all the variables in Table 1. Stepwise backward variable selection process was used for this analysis using cutoff of p-value of less than 0.10. Nagelkerke’s R2 was 0.32 and Hosmer-Lemeshow goodness-of-fit test showed good calibration (chi-square = 12.1, p = 0.231). CI = confidence interval, E/e´ = ratio of early transmitral flow velocity to early diastolic velocity of the mitral annulus.

Simple decision tree model showing the classification of patients with and without AKI is shown in Figure 2. The importance matrix plot of gradient boosting is shown in Figure 3 and the amount of Intraoperative red blood cells transfusion and preoperative hematocrit level were ranked the first and second. The variables of importance plot of random forest model was shown in Figure S1. The same variables were ranked first and second in terms of both mean decreases in accuracy and Gini. The matrix of classification of extreme gradient boosting was visualized in Figure S2. Figure S3 shows an example of the support vector machine classification plot.
Figure 2

Simple decision tree model showing the classification of patients with (1) and without (0) acute kidney injury (AKI). The numbers with two decimals in each cell means the probability of developing AKI in each classification tree. The blue or green color becomes dense when it is more likely to develop acute kidney injury or not. The % number in the boxes denotes the percentage of patients with each discriminating variable from CART (Classification And Regression Tree) analysis. Intraop = intraoperative, preop = preoperative, pRBC = packed red blood cells, Hct = hematocrit, Cr = creatinine, FFP = fresh frozen plasma, E_or_e_prime = preoperative ratio of early transmitral flow velocity to early diastolic velocity of the mitral annulus.

Figure 3

Importance matrix plot of the gradient boosting machine. This figure shows the importance of each covariates in the final model. ARB = angiotensin receptor blocker, BMI = body-mass index, CABG = coronary artery bypass graft, CCB = calcium channel blocker, CKD = chronic kidney disease, Cr = creatinine, CVA = history of cerebrovascular accident, EF = ejection fraction, E_or_e_prime = preoperative ratio of early transmitral flow velocity to early diastolic velocity of the mitral annulus, FFP = fresh frozen plasma, hct = hematocrit, HTN = hypertension, intraop = intraoperative, mean SvO2 = intraoperative mean mixed venous oxygen saturation, three_VD = three vessel coronary disease, preop = preoperative, pRBC = packed red blood cells.

4. Discussion

We compared the predictive accuracy of the prediction for AKI after cardiovascular surgery among the machine learning techniques, traditional statistical approach, and previous risk scoring models. We included currently available machine learning techniques, including decision tree, random forest, support vector machines, neural networks, and deep belief networks. Logistic regression analysis was used as the traditional approach. The results showed that extreme gradient boosting machine showed the lowest error rate and largest AUC among all techniques and risk scores, which was consistent for the prediction of stage 2 or 3 AKI. Extreme gradient boosting machine based prediction may result in significant improvement in the prediction of AKI after cardiac surgery. A risk estimator based on our gradient boosting model was developed for clinical use to determine the risk of AKI at the end of surgery. Extreme gradient boosting showed the best predictive ability in our analysis [32,33,37]. While the random forest builds an ensemble of independent recursive partitioning tress of unlimited depth, extreme gradient boosting builds a sequential series of shallow trees, where each tree corrects for the residuals in the predictions made by all the previous tress (Figure S2). Gradient boosting uses techniques to reduce overfitting such as shrinkage and column resampling. After each step of boosting, the algorithm scales the newly added weights, which reduces the influence of each tree and allowing the model to learn better. Column resampling considers only a random subset of descriptors in building a given tree, which also fastens the training process by reducing the number of descriptors to consider [32]. It may be determined in further multicenter larger studies whether the better performance of boosting could be applied to data of other institutions or other surgical populations. Decision tree analysis showed a similar performance to that of logistic regression model in our study. Decision trees are a hierarchical model that are comprised of decision rules based on the optimal feature cutoff values. It recursively classifies independent variables into different small groups based on the Gini impurity measure or entropy, while logistic regression analysis analyzes the interaction of included variables [38,39,40]. The odds ratio of a specific risk factor in a logistic regression model is applied to all study population rather than a single subgroup, while each branch of the decision tree may have different covariates from another branch. Variable selection in the process of decision tree is not based on probabilistic methods, which may result in overestimation of the importance of explanatory variables or may miss other potential confounders [41]. Decision trees can improve the predictive ability achieved by logistic regression models under certain circumstances. With sufficiently many terminal nodes with a low coefficient of variance, the decision tree model enables the detection of some individual cases that would have been unnoticed applying conventional logistic regression models. However, the clinical interpretation of variable selection and their cutoffs is often difficult, because the decision tree classification does not consider the clinical relevance. Decision trees are susceptible to fluctuations in the training set and are, thus, prone to overfitting and poor generalizability [4]. Additionally, decision tree models may not be practically useful if it includes too many variables. However, for the low error rate and high AUC, more classifying variables are needed. The performance of random forest was also similar to that of logistic regression analysis in our dataset. Random forest is considered to have advantages, especially in handling electronic medical records. It is an extension to traditional decision tree classifiers [42], and attempts to mitigate the limitations of decision tree through an ensemble-based technique using multiple decision trees. Each tree is constructed from a random subset of the original training data and a random subset of a total number of variables is analyzed at each node for splitting. Random forests can minimize the problem of overfitting by taking the mode of decisions of a large number of these randomly generated trees [43]. Other advantages of random forests to analyze electronic medical records include running efficiently on large samples with thousands of input variables, the ability to accommodate different data scales, and robustness to the inclusion of irrelevant variables. There was no significant performance gain of the random forest over that of the simple decision tree in our study, which may be because the number of input variables was insufficient to demonstrate any difference in performance. The deep neural network model showed a good performance to predict in-hospital mortality in a previous study, although it was not superior to previous risk score [5]. Contrary to our expectations, the performance of neural network in our study was inferior to the performances of all other machine learning techniques. This may be explained because our data for learning the relationship between the covariates and risk of AKI may not be sufficient. Although the multilayer perceptron is mathematically proven to be able to approximate any nonlinear function, it requires a large amount of learning data. Therefore, the dataset of our study may not be large enough and the number of covariates was not sufficient to train the multilayer perceptron [44]. The performance of previous eight risk scoring models was poor in our test dataset [9,10,11,12,13,14,15,16]. The AUCs of these risk scores were similar possibly because similar predictors were used to construct the risk score [6], and the poor performance may be due to the small number of predictors and lack of intraoperative variables, such as transfusion amounts or hemodynamic variables. A previous study showed that the performance of the logistic regression model could be improved when we consider many perioperative variables as possible [19]. Several previous studies reported that the AUCs of machine learning techniques were not superior to previous risk scores or logistic regression models to predict postoperative mortality [5,45]. However, our study demonstrated that the AUCs of machine learning techniques could be significantly greater than the AUC of logistic regression model to predict AKI. Previous studies compared the predictive ability for in-hospital mortality in a population with a very low incidence (<1%) [5,45]. The difference in AUC or error rate may be small for an outcome with low incidence, and this small difference in performance would be difficult to be demonstrated. It seems that any difference in error rate and AUC would be more pronounced in our study sample with a postoperative AKI of higher incidence (38.3%). This could also be the reason why the SMOTE model of random forests or support vector machines did not significantly increase the AUC in our test dataset. SMOTE model increases the incidence of outcome cases and balancing the case with and without outcome variables. However, our test set already had a nearly balanced dataset for AKI. The importance matrix plot of the gradient boosting machine shows the similar predictors that were known to be associated with the development of AKI after cardiac surgery [6,9,10,11,12,13,14,16,19,20,21,46]. However, the plot additionally gives the relative importance of each predictor, which was similar to the variance importance plot of random forest model. This analysis may help to find a new risk factor for postoperative morbidity or mortality. Our study has several limitations. First, our analysis used only single-center data and included a relatively small number of cases and covariates. The performance of machine learning techniques might be different when they are applied to a much larger sample with a different distribution of the covariates. The external validity of our results may be limited. Furthermore, important predictors may be different according to different institutions. However, the relative performance of logistic regression and machine learning techniques would be similar to our results. Each institution may need to develop their own prediction model with the machine learning approach, by using historical data from their electronic medical records and updating the model periodically. Real-time processing of patient data would produce risk prediction for each patient after surgery. Second, machine learning techniques are often difficult to interpret the results. Inferences about the explanatory variables are more difficult than logistic regression analysis [4]. However, the gradient boosting machine and random forest provided for some interpretability through the importance matrix plot and variable importance plot. Third, it is not certain that our results could translate into improved clinical outcomes for the patients. Most of our important variables reported are not clinically modifiable and accurate risk prediction may not be followed by improved patient outcomes. However, further prospective trials may evaluate whether adjustment of potentially modifiable predictors, such as hemodynamic variables could decrease the risk of AKI [46,47,48].

5. Conclusions

In conclusion, our study demonstrated that the machine learning technique of extreme gradient boosting showed significantly better performance than the traditional logistic regression analysis or previous risk scores in predicting both AKI of all stages and stage 2 or 3 AKI after cardiac surgery. Gradient boosting machine may be used for real-time processing of patient data to estimate the risk of AKI after cardiac surgery at the end of surgery. Our Internet-based risk estimator may help to evaluate the risk of AKI at the end of surgery. However, prospective multicenter trials are required to validate the better prediction by gradient boosting. Further studies may apply extreme gradient boosting machine to the other important clinical outcomes after cardiac surgeries and may prospectively validate our results.
  44 in total

1.  A tree-based decision rule for identifying profile groups of cases without predefined classes: application in diffuse large B-cell lymphomas.

Authors:  Elias Zintzaras; Maria Bai; Christos Douligeris; Axel Kowald; Panayiotis Kanavaros
Journal:  Comput Biol Med       Date:  2006-08-08       Impact factor: 4.589

2.  Intraoperative oliguria predicts acute kidney injury after major abdominal surgery.

Authors:  T Mizota; Y Yamamoto; M Hamada; S Matsukawa; S Shimizu; S Kai
Journal:  Br J Anaesth       Date:  2017-12-01       Impact factor: 9.166

3.  Modern hydroxyethyl starch and acute kidney injury after cardiac surgery: a prospective multicentre cohort.

Authors:  M Vives; R Callejas; P Duque; G Echarri; D N Wijeysundera; A Hernandez; A Sabate; M Bes-Rastrollo; P Monedero
Journal:  Br J Anaesth       Date:  2016-10-17       Impact factor: 9.166

4.  Predicting acute renal failure after coronary bypass surgery: cross-validation of two risk-stratification algorithms.

Authors:  E B Fortescue; D W Bates; G M Chertow
Journal:  Kidney Int       Date:  2000-06       Impact factor: 10.612

5.  Logistic risk model predicting postoperative renal failure requiring dialysis in cardiac surgery patients.

Authors:  Parwis B Rahmanian; Grzegorz Kwiecien; Georg Langebartels; Navid Madershahian; Thorsten Wittwer; Thorsten Wahlers
Journal:  Eur J Cardiothorac Surg       Date:  2011-02-21       Impact factor: 4.191

6.  Association Between Elevated Echocardiographic Index of Left Ventricular Filling Pressure and Acute Kidney Injury After Off-Pump Coronary Artery Surgery.

Authors:  Min Hur; Karam Nam; Woo Young Jo; Gahyun Kim; Won Ho Kim; Jae-Hyon Bahk
Journal:  Circ J       Date:  2017-12-05       Impact factor: 2.993

7.  ROSE: decision trees, automatic learning and their applications in cardiac medicine.

Authors:  J Zavrsnik; P Kokol; I Malèiae; K Kancler; M Mernik; M Bigec
Journal:  Medinfo       Date:  1995

8.  Acute kidney injury prediction following elective cardiac surgery: AKICS Score.

Authors:  H Palomba; I de Castro; A L C Neto; S Lage; L Yu
Journal:  Kidney Int       Date:  2007-07-11       Impact factor: 10.612

9.  Defining oliguria during cardiopulmonary bypass and its relationship with cardiac surgery-associated acute kidney injury.

Authors:  D Hori; N M Katz; D M Fine; M Ono; V M Barodka; L C Lester; G Yenokyan; C W Hogue
Journal:  Br J Anaesth       Date:  2016-12       Impact factor: 9.166

10.  Potentially modifiable risk factors for acute kidney injury after surgery on the thoracic aorta: a propensity score matched case-control study.

Authors:  Won Ho Kim; Mi Hye Park; Hyo-Jin Kim; Hyun-Young Lim; Haeng Seon Shim; Ju-Tae Sohn; Chung Su Kim; Sangmin M Lee
Journal:  Medicine (Baltimore)       Date:  2015-01       Impact factor: 1.889

View more
  35 in total

Review 1.  Artificial intelligence-enabled decision support in nephrology.

Authors:  Tyler J Loftus; Benjamin Shickel; Tezcan Ozrazgat-Baslanti; Yuanfang Ren; Benjamin S Glicksberg; Jie Cao; Karandeep Singh; Lili Chan; Girish N Nadkarni; Azra Bihorac
Journal:  Nat Rev Nephrol       Date:  2022-04-22       Impact factor: 42.439

2.  Artificial intelligence in cardiothoracic surgery.

Authors:  Roger D Dias; Julie A Shah; Marco A Zenati
Journal:  Minerva Cardioangiol       Date:  2020-09-29       Impact factor: 1.347

3.  Predictive Model for High Coronary Artery Calcium Score in Young Patients with Non-Dialysis Chronic Kidney Disease.

Authors:  Tae Ryom Oh; Su Hyun Song; Hong Sang Choi; Sang Heon Suh; Chang Seong Kim; Ji Yong Jung; Kyu Hun Choi; Kook-Hwan Oh; Seong Kwon Ma; Eun Hui Bae; Soo Wan Kim
Journal:  J Pers Med       Date:  2021-12-15

Review 4.  Artificial intelligence in perioperative medicine: a narrative review.

Authors:  Hyun-Kyu Yoon; Hyun-Lim Yang; Chul-Woo Jung; Hyung-Chul Lee
Journal:  Korean J Anesthesiol       Date:  2022-03-29

5.  Prediction of Acute Kidney Injury after Liver Transplantation: Machine Learning Approaches vs. Logistic Regression Model.

Authors:  Hyung-Chul Lee; Soo Bin Yoon; Seong-Mi Yang; Won Ho Kim; Ho-Geol Ryu; Chul-Woo Jung; Kyung-Suk Suh; Kook Hyun Lee
Journal:  J Clin Med       Date:  2018-11-08       Impact factor: 4.241

6.  Machine Learning Models of Survival Prediction in Trauma Patients.

Authors:  Cheng-Shyuan Rau; Shao-Chun Wu; Jung-Fang Chuang; Chun-Ying Huang; Hang-Tsung Liu; Peng-Chen Chien; Ching-Hua Hsieh
Journal:  J Clin Med       Date:  2019-06-05       Impact factor: 4.241

Review 7.  Role of Artificial Intelligence in Kidney Disease.

Authors:  Qiongjing Yuan; Haixia Zhang; Tianci Deng; Shumei Tang; Xiangning Yuan; Wenbin Tang; Yanyun Xie; Huipeng Ge; Xiufen Wang; Qiaoling Zhou; Xiangcheng Xiao
Journal:  Int J Med Sci       Date:  2020-04-06       Impact factor: 3.738

Review 8.  Machine learning in nephrology: scratching the surface.

Authors:  Qi Li; Qiu-Ling Fan; Qiu-Xia Han; Wen-Jia Geng; Huan-Huan Zhao; Xiao-Nan Ding; Jing-Yao Yan; Han-Yu Zhu
Journal:  Chin Med J (Engl)       Date:  2020-03-20       Impact factor: 2.628

9.  Diagnostics, Risk Factors, Treatment and Outcomes of Acute Kidney Injury in a New Paradigm.

Authors:  Charat Thongprayoon; Panupong Hansrivijit; Karthik Kovvuru; Swetha R Kanduri; Aldo Torres-Ortiz; Prakrati Acharya; Maria L Gonzalez-Suarez; Wisit Kaewput; Tarun Bathini; Wisit Cheungpasitporn
Journal:  J Clin Med       Date:  2020-04-13       Impact factor: 4.241

10.  Prediction of the development of acute kidney injury following cardiac surgery by machine learning.

Authors:  Po-Yu Tseng; Yi-Ting Chen; Chuen-Heng Wang; Kuan-Ming Chiu; Yu-Sen Peng; Shih-Ping Hsu; Kang-Lung Chen; Chih-Yu Yang; Oscar Kuang-Sheng Lee
Journal:  Crit Care       Date:  2020-07-31       Impact factor: 9.097

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.