Literature DB >> 30830157

Towards the best kidney failure prediction tool: a systematic review and selection aid.

Chava L Ramspek1, Ype de Jong1,2, Friedo W Dekker1, Merel van Diepen1.   

Abstract

BACKGROUND: Prediction tools that identify chronic kidney disease (CKD) patients at a high risk of developing kidney failure have the potential for great clinical value, but limited uptake. The aim of the current study is to systematically review all available models predicting kidney failure in CKD patients, organize empirical evidence on their validity and ultimately provide guidance in the interpretation and uptake of these tools.
METHODS: PubMed and EMBASE were searched for relevant articles. Titles, abstracts and full-text articles were sequentially screened for inclusion by two independent researchers. Data on study design, model development and performance were extracted. The risk of bias and clinical usefulness were assessed and combined in order to provide recommendations on which models to use.
RESULTS: Of 2183 screened studies, a total of 42 studies were included in the current review. Most studies showed high discriminatory capacity and the included predictors had large overlap. Overall, the risk of bias was high. Slightly less than half the studies (48%) presented enough detail for the use of their prediction tool in practice and few models were externally validated.
CONCLUSIONS: The current systematic review may be used as a tool to select the most appropriate and robust prognostic model for various settings. Although some models showed great potential, many lacked clinical relevance due to being developed in a prevalent patient population with a wide range of disease severity. Future research efforts should focus on external validation and impact assessment in clinically relevant patient populations.
© The Author(s) 2019. Published by Oxford University Press on behalf of ERA-EDTA.

Entities:  

Keywords:  kidney failure; prediction model; prognostic; systematic review

Mesh:

Year:  2020        PMID: 30830157      PMCID: PMC7473808          DOI: 10.1093/ndt/gfz018

Source DB:  PubMed          Journal:  Nephrol Dial Transplant        ISSN: 0931-0509            Impact factor:   5.992


BACKGROUND

Chronic kidney disease (CKD) may lead to kidney failure, although the rates of progression vary substantially between individuals [1]. Prediction tools that can identify patients at high risk of developing kidney failure could have great clinical value. They could be used to inform individualized decision making, employed in determining the appropriate time for referral to nephrologists and used in the planning and preparation of renal replacement therapy (RRT). Prediction tools might also offer opportunities for risk stratification in research and improvement of health policies [2]. Multiple prediction models have been developed to identify individuals at high risk of kidney failure and have been previously described in two systematic reviews [3, 4]. Many of these models showed good predictive abilities in development. However, despite nephrologists and patients acknowledging a lack of prognosis discussions in practice, clinical uptake of these tools is still limited [5]. Policymakers also seem hesitant in endorsing prediction tools. The most recent Kidney Disease: Improving Global Outcomes guideline recommends the use of prediction models for timely referral for planning RRT [6]. However, the guideline, fails to provide guidance on which risk prediction tool should be used to do so. The lack of uptake by clinicians and policymakers has been partly attributed to substandard methodology, a lack of external validation and a shortage of easy calculation options [7]. The last two published reviews in 2012 and 2013 included eight studies each on prediction of kidney failure in CKD patients [3, 4]. Since then the number of available models has greatly increased. A new systematic review of the available models is the first step towards the use and recommendation of robust prognostic tools. The aim of the current study is therefore to systematically review all available models predicting kidney failure in CKD patients, organize empirical evidence on their validity and ultimately provide guidance in the selection of the best prediction tool for various settings.

MATERIALS AND METHODS

Data sources and searches

The current review was framed by the search for prognostic prediction models for CKD patients, predicting the future event of kidney failure. To ensure transparent reporting and accurate study appraisal, the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA), Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD) and Checklist for Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modelling Studies (CHARMS) guidelines were followed where applicable [8-10]. The completed PRISMA checklist is provided as supplementary material. We searched the PubMed and EMBASE databases on 31 December 2017 for English-language studies regarding risk prediction in CKD patients. The search strategies were designed to include relevant development, validation and implementation studies and are provided in Appendix A1.

Study selection

Titles, abstracts and full-text articles were sequentially screened for inclusion by two independent researchers (C.L.R. and Y.J.). Discrepancies on inclusion of full-text articles were resolved by consulting a third co-author (M.D.). Articles were included if they met the following predefined selection criteria: (i) the study must develop, validate, update or implement a multivariate prognostic prediction model, with a prediction research question as the aim, as opposed to an aetiological or methodological goal; (ii) the study must present at least one measure to assess model performance; (iii) the study population must consist of adult CKD patients and (iv) the study outcome must include kidney failure or end-stage renal disease. The references of included studies and related reviews were manually screened in order to identify additional relevant studies.

Data extraction and quality assessment

Following selection, two reviewers (C.L.R. and Y.J.) independently conducted the data extraction and quality assessment. Discrepancies were discussed with input from an additional co-author (M.D.) where necessary. Conforming with CHARMS recommendations, information on the source of the data, population, outcome, sample size, missing data, model development and model performance were extracted and summarized. Additionally, data on external validations of models were extracted. Furthermore, the risk of bias and clinical usefulness were judged by both reviewers independently. In order to facilitate further comparison, studies were grouped by study population, which ranged from very broad (general CKD) to specific CKD subgroups such as immunoglobulin A (IgA) nephropathy or diabetic nephropathy. Quality and risk of bias were assessed in both development and validation studies by making use of a novel tool, the Prediction Study Risk of Bias Assessment Tool (PROBAST). Although this tool has yet to be published in its complete form, there is no other formal risk of bias assessment available that is applicable to prediction studies. The PROBAST is specifically designed for systematic reviews of prediction studies and is used as a domain-based approach with 23 signalling questions that categorize the risk of bias into high, low or unclear for five separate domains: participant selection, predictors, outcome, sample size and missing data, and analysis. It also assesses the usability of a model. It has been used in multiple reviews in the past year and was presented in part at the 2016 Cochrane Colloquia [11]. The final test version of PROBAST was obtained through personal e-mail contact with Dr R.C.G. Wolff.

Data synthesis

Given the multitude of different models and heterogeneity in study characteristics, we opted for a narrative synthesis of results supported by extensive tables and figures with study characteristics listed per article. Model performance was evaluated by examining the discrimination and calibration of included prediction tools. Discrimination is most often described by the C-statistic and indicates how well the model discriminates between patients with and without the event of interest. It lies between 0.5 and 1, where 0.5 is similar to tossing a coin and 1 indicates perfect discrimination [12]. Important to take into account is that the C-statistic of the same model can vary greatly, dependent on the population on which the model is tested. When a population is heterogeneous in the predictors that make up the prediction tool, the C-statistic may increase substantially [13]. On the other hand, calibration describes the agreement between the absolute number of predicted events and observed events population wide. It is best represented in a plot, wherein the predicted probability of kidney failure is plotted against the observed rate of kidney failure [12]. To evaluate the sample size and risk of overfitting in development studies, the events per candidate predictor (EPV) were extracted. A minimum of 10 EPV has been suggested as rule of thumb for an acceptable sample size in model development studies [14]. For external validation studies, it has been recommended to include a minimum of 100 events in total to obtain a precise estimate of performance [15].

RESULTS

The study selection process is described in a flowchart (Figure 1). Overall, 2183 titles were identified, of which 431 abstracts were assessed, and 90 full-text publications were evaluated in depth. From these articles, a final 42 studies met all inclusion criteria and were included in the current review. Most full-text exclusions were due to the predicted outcome not including kidney failure or the lack of a multivariate model. Although prediction research has seen a great surge in nephrology over the last few years, the first included predictive model was published in 1986 for IgA nephropathy patients. Since the beginning of the 2000s, a substantial increase of published models is apparent, as can be seen in Figure 2. Although the number of developed models has increased almost every year, the number of validation studies has remained small. Of the 42 included studies, 7 exclusively externally validated already existing models [16-22]. Besides development, 10 studies also externally validated their own or previously published models. Disconcertingly, no study assessing the impact of using such a prediction tool was found, which ultimately is the only way of assessing whether the model can improve patient care.
FIGURE 1

PRISMA flow diagram of study inclusion.

FIGURE 2

Cumulative number of published development and validation studies for models that predict kidney failure in CKD patients (N = 42).

PRISMA flow diagram of study inclusion. Cumulative number of published development and validation studies for models that predict kidney failure in CKD patients (N = 42).

Characteristics of development studies

A total of 35 studies were published on the development of novel tools to predict kidney failure in CKD patients. Generally, a distinction can be made between models developed for a general CKD patient population (n = 16) and models developed for a population with a specified primary renal disease (n = 19), mainly IgA nephropathy or diabetic nephropathy. The characteristics of all included development studies are described in Table 1. Since each study developed between 1 and 12 prediction models, the results presented in Table 1 concern the final model(s) as selected by the authors or the model with the best performance if no final model was suggested. The population size differed greatly between studies and ranged from 75 to 28 779 patients. A small sample size was a problem in 17/35 studies, as they had <10 EPV, thus running the substantial risk of overfitting their model [14]. To assess to what extent these models are overfit, external validation is of key importance. Before the validity of these models is tested, they should not be used in practice.
Table 1

Baseline characteristics of model development studies (N = 35)

StudyCountryDesignPopulationMean GFR N total, No. of eventsOutcomeTime frame (years)EPVModel type N predictorsInternal validationC-statisticCalibrationPresented model
General CKD
Cheng et al. [23]TaiwanSingle-centre cohortGeneral CKD Stage 4463, 132GFR <150.53CART11Cross-validationD: 0.72Decision rules
Schroeder et al. [24]USAMulticentre cohortGeneral CKD Stages 3 and 4a4722 460, 737RRT574Cox8Bootstrap (+external)IV: 0.96D: plotFormulab and score
Hsu et al. [25]USACohortGeneral CKD GFR 20–70442466, 581RRT, 50% GFR ↓36Cox12D: 0.89HRs
Tangri et al. [26] AJKDCanadaSingle-centre cohortGeneral CKD Stages 3–5363004, 344RRTDynamic43Cox8Bootstrap, cross- validationIV: 0.91D: plot, testFormula
Xie et al. [27]USAMulticentre cohortGeneral CKD Stages 3–5c4928 779, 1730RRT1, 3, 5, 7115Cox5Cross-validationIV: 0.92HRs
Marks et al. [28]ScotllandMulticentre cohortGeneral CKD Stages 3–5333396, 142RRT524Logistic5– (external)D: 0.94D: testFormula
Maziarz et al. [29]USAMulticentre cohortGeneral CKD Stages 3–5c28 779, 1730RRT1, 3, 5115Cox5Cross-validationIV: 0.92HRs
Levin et al. [30]CanadaMulticentre cohortGeneral CKD Stages 3–4282402, 142RRT19Cox7BootstrapD: 0.87D: testHRs
Maziarz et al. [31]USAMulticentre cohortGeneral CKD Stages 3–5c16 656, 959RRT1, 3, 563Cox5Cross-validationIV: 0.90
Drawz et al. [32]USASingle-centre cohortGeneral CKD Stages 4 and 5d251866, 77RRT14Cox6Bootstrap (+external)IV: 0.86Formula
Smith et al. [33]UKMulticentre cohortGeneral CKD Stages 3 and 432158, 40Death, RRT24Cox10D: 0.81HRs
Tangri et al. [34]CanadaSingle-centre cohortGeneral CKD Stages 3–5363449, 386RRT1, 3, 516Cox4, 8– (external)4v: 0.91 8v: 0.92Formula and web calculator
Landray et al. [35]UKSingle-centre cohortGeneral CKD Stages 3–522382, 190RRT4Cox4– (external)D: 0.87D: plotHRs
Johnson et al. [36]USAMulticentre cohortGeneral CKD Stages 3 and 4a9782, 323RRT554Cox6BootstrapIV: 0.89D: plotScore
Johnson et al. [37]USAMulticentre cohortGeneral CKD Stages 3–5a6541, 369RRT541Cox6D: 0.91HRs
Dimitrov et al. [38]ItalyRCTGeneral CKD GFR 20–7043344, 80ESRD7ANN4D: 0.80Decision tree
Specified renal disease
Bidadkosh et al. [39]MultinationalRCTDiabetic nephropathy33861, 60ESRD, 40% GFR ↓6Cox8D: 0.79
Tang et al. [40]ChinaSingle-centre cohortLupus nephritis78599, 145RRT, 50% GFR ↓, GFR <154Cox8Split sampleHRs and score
Barbour et al. [41]MultinationalMulticentre cohortIgA nephropathy68901, 162GFR <15, 50% GFR↓521Cox8BootstrapD: 0.80D: plotFormula
Li et al. [42]TaiwanSingle-centre cohortDiabetic nephropathy131, 22RRT2Cox4Cross-validationD: 0.90HRs and score
Pesce et al. [43]MultinationalMulticentre cohortIgA nephropathy871040, 241Time to ESRD3–824ANN6Split sample + cross- validationIV: 0.90Web calculator (out of service)
Diciolla et al. [44]MultinationalMulticentre cohortIgA nephropathy1040, 241RRT540ANN6Cross-validationWeb calculator (out of service)
Hoshino et al. [45]JapanSingle-centre cohortDiabetic nephropathy44205, –RRT10Cox4Cross-validationIV: 0.93
Tanaka et al. [46]JapanMulticentre cohortIgA nephropathy698, 73RRT57Cox5– (external)D: 0.87D: plot, testHRs and score
Xie et al. [47]ChinaSingle-centre cohortIgA nephropathy88619, 67ESRD2, 5, 102Cox4D: 0.85HRs
Berthoux et al. [48]FranceSingle-centre cohortIgA nephropathy75332, 45Death, RRT10, 208Score3– (external)HRs and score
Desai et al. [49]MultinationalMulticentre RCTDiabetic nephropathy35995, 222RRT6Cox19BootstrapD: 0.85HRs
Day et al. [50]UKSingle-centre cohortPauci-immune GN390, 54RRT19Cox2D: 0.83HRs
Goto et al. [51]JapanMulticentre cohortIgA nephropathye2283, 252RRT1018Cox8Bootstrap + split-sampleD: 0.94 IV: 0.94Score
Kent et al. [52]MultinationalMultiple RCT’sNon-diabetic CKD1860, 311RRT/100% SCr ↑62Cox5D: 0.83D: plot, testHRs
Keane et al. [53]MultinationalRCTDiabetic nephropathy1513, 341RRT12Cox4JackknifeD: plotHRs
Magistroni et al. [54]ItalySingle-centre cohortIgA nephropathy83237, 40RRT102Cox4– (external)Score
Wakai et al. [55]JapanMulticentre cohortIgA nephropathye2269, 207ESRD719Cox8Bootstrap + split-sampleD: 0.94 IV: 0.93Score
Frimat et al. [56]FranceSingle-centre cohortIgA nephropathy210, 33RRT72Cox7D: plotScore
Beukhof et al. [57]The NetherlandsSingle-centre cohortIgA nephropathy9475, 14RRT101Cox5Nomogram

Both studies by Johnson et al. [36, 37] overlap in patient population and include the same predictors. The study by Schroeder et al. [24] updates this same model [36] (the KPNW) by including additional predictors and excluding some original predictors. bThe formula as provided in the supplement of Schroeder et al.’s article [24] does not provide the knot locations for spline terms. These are available from the authors upon request. cThe study by Xie [27] and Maziarz [29] includes the same patient population. Part of this population is included in Maziarz et al. [31]. All three studies include the same predictors in the same four models but re-estimate β-coefficients for different subsets. The population is of underserved/uninsured patients. dPopulation of veterans ≥65 years old. eOverlap in patient population. The study by Goto et al. [51] has an extended follow-up of 3 years in the same cohort as the study by Wakai et al. [55]. “–” not reported. (e)GFR, (estimated) glomerular filtration rate in mL/min/1.73 m2; EPV, events per variable/candidate predictor; D, development; IV, internal validation; CART, classification and regression tree; ANN, artificial neural networks; RCT, randomized control trial; SCr, serum creatinine.

Baseline characteristics of model development studies (N = 35) Both studies by Johnson et al. [36, 37] overlap in patient population and include the same predictors. The study by Schroeder et al. [24] updates this same model [36] (the KPNW) by including additional predictors and excluding some original predictors. bThe formula as provided in the supplement of Schroeder et al.’s article [24] does not provide the knot locations for spline terms. These are available from the authors upon request. cThe study by Xie [27] and Maziarz [29] includes the same patient population. Part of this population is included in Maziarz et al. [31]. All three studies include the same predictors in the same four models but re-estimate β-coefficients for different subsets. The population is of underserved/uninsured patients. dPopulation of veterans ≥65 years old. eOverlap in patient population. The study by Goto et al. [51] has an extended follow-up of 3 years in the same cohort as the study by Wakai et al. [55]. “–” not reported. (e)GFR, (estimated) glomerular filtration rate in mL/min/1.73 m2; EPV, events per variable/candidate predictor; D, development; IV, internal validation; CART, classification and regression tree; ANN, artificial neural networks; RCT, randomized control trial; SCr, serum creatinine. For specific renal diseases, the baseline was almost always the first biopsy (and disease confirmation), providing a clear moment in time for when to use the prognostic model or score. Models developed in general CKD, however, rarely defined the moment in time when their prediction tool should be used, as most of these studies enrolled prevalent CKD patients with a wide range of disease severity. Only two models were developed on incident patients, who were included at the first referral to a nephrologist [26, 34]. There was some variation in outcome definitions, but for most studies, renal failure was defined as the need for RRT (dialysis start or kidney transplantation). Five studies used estimated glomerular filtration rate (eGFR) or creatinine as a proxy for kidney failure. Two development studies used RRT start or death as a composite outcome measure. A total of four studies did not report their definition of ESRD. The time frame over which the models predict kidney failure ranged from 6 months to 20 years and nine studies failed to define a prediction time frame, presumably using the maximum study follow-up. The specific predictors included per development study are presented in Figure 3. There is a large amount of overlap in final predictors with almost all studies including age, sex, eGFR (or serum creatinine), proteinuria and histological features for IgA nephropathy tools.
FIGURE 3

Predictors included in development studies (N = 35). The inclusion of a predictor is shown as ‘X’. The subscript under X (e.g. ‘X2’) indicates the number of predictors included from that category.

Predictors included in development studies (N = 35). The inclusion of a predictor is shown as ‘X’. The subscript under X (e.g. ‘X2’) indicates the number of predictors included from that category. Concerning the reporting of performance measures, discrimination measures were reported far more often than calibration measures. Discrimination in the form of a C-statistic was reported in 28/35 studies. The C-statistic ranged from 0.72 to 0.96 and was generally high, indicating good to excellent discrimination in most studies. Calibration was presented far less frequently, with only 11 studies presenting a calibration plot, bar chart or test. In order to calculate an individual’s risk, the model constant and hazard ratios (HRs)/regression coefficients per predictor are needed. Many studies only presented HRs per predictor without the constant (intercept or baseline hazard value), and some gave no data on the model equation at all. The full formula for the developed model was presented in only 6/35 studies. Just three studies provided a web calculator for easy use, of which two web calculators are no longer working. A total of 13 studies provided a simplified scoring system. In total, 25 final models were validated in some form, either internally and/or externally. Cross-validation, bootstrapping and random split sample were the most used forms of internal validation.

Characteristics of external validation studies

A total of 17 studies externally validated one or more of the developed prediction tools. The characteristics of these models and validations can be found in Table 2. Most validation studies were performed by the same group of researchers who developed the models and were often presented in the same publication as the development. Compared with the development performance, the C-statistic was lower in 68% of the validations. Two studies updated the validated model by recalibrating the baseline hazard and two studies added predictors to the existing model. In total, five risk scores predicting prognosis in IgA nephropathy patients and seven prognostic tools for general CKD patients were externally validated. Only the Absolute Renal Risk (ARR) score, Goto score and Kidney Failure Risk Equation (KFRE) (three, four and eight variables) were validated multiple times. The largest validation study of the KFRE was performed by Tangri et al. [18] and summarized the validation of the KFRE in >30 countries, including more than half a million patients.
Table 2

Characteristics of external validation studies and model performance in validations (N = 17)

StudyModel validatedIndependentValidation typeCountryPopulationGFR mean N total, No. of eventsOutcomeTime frame (years)Model updatedC-statisticCalibrationUpdated model presented
General CKD
Schroeder et al. [24]KPNW model (Johnson)NoExternalUSAGeneral CKD Stages 3 and 44816 553, 360RRT5Baseline hazard recalibrated to Colorado0.95PlotNo
Lennartz et al. [19]KFRE 4v (Tangri)NoExternalGermanyGeneral CKD Stages 2–446565, 52RRT3Baseline hazard, addition ultrasound parameters4v, update: 0.91, 0.91PlotFormula
Tangri et al. [18]KFRE 4v and 8v (Tangri)NoExternal>30 countriesGeneral CKD Stages 3–546721 357, 23 829RRT2, 5Baseline hazard recalibrated to Europe4v, 8v: 0.88, 0.88PlotFormula and web calculator
Grams et al. [16]KFRE 4v (Tangri)YesExternalUSACKD, GFR 20–65a1094, –RRT1, 5No0.83
Marks et al. [28]Model 7 (Marks)NoTemporalScotlandGeneral CKD Stages 3–54718 687, 222RRT5No0.96Plot, test
KFRE 3v and 4v (Tangri)YesExternalScotlandGeneral CKD Stages 3–54718 687, 222RRT5No3v, 4v: 0.94, 0.95Plot
Levin et al. [30]KFRE 8v (Tangri)NoExternalCanadaGeneral CKD Stages 3b–4282402, 142RRT1Coefficients re-estimated, biomarkers added8v, update: 0.86, 0.87HRs
Drawz et al. [32]VA risk score (Drawz)NoGeographicUSAGeneral CKD Stage 4 and 5b25819, 33GFR <15, RRT1No0.82
KFRE 8v (Tangri)YesExternalUSAGeneral CKD Stage 4 and 5b252684, 110GFR <15, RRT1No0.78
Peeters et al. [17]KFRE 3v, 4v, and 8v (Tangri)YesExternalThe Neth- erlandsGeneral CKD Stages 3–533595, 114RRT5No3v, 4v, 8v: 0.88, 0.88, 0.89Plot, test
Tangri et al. [34]KFRE 3v, 4v, and 8v (Tangri)NoExternalCanadaGeneral CKD Stages 3–5314942, 1177RRT1, 3, 5No3v, 4v, 8v: 0.79, 0.83, 0.84Plot, test
Landray et al. [35]CRIB score (Landray)NoExternalUKGeneral CKD Stages 3–522213, 66RRTNo0.91Plot
Specified renal disease
Knoop et al. [21]ARR score (Berthoux)YesExternalNorwayIgA nephropathy1134, 320Death, RRT5, 10, 15Coefficients re-estimated, age and GFR added0.79 update: 0.89Formula
Mohey et al. [22]ARR score (Berthoux)NoExternalFranceSecondary IgA nephropathy8274, 19GFR <15, death10, 20No
Tanaka et al. [46]Tanaka scoreNoExternalJapanIgA nephropathy702, 85RRT5No0.89Plot, test
Xie et al. [47]Goto scoreYesExternalChinaIgA nephropathy88619, 67ESRD2, 5, 10No0.82
RENAAL score (Keane)YesExternalChinaIgA nephropathy88619, 67ESRD2, 5, 10No0.79
ARR score (Berthoux)YesExternalChinaIgA nephropathy88619, 67ESRD2, 5, 10No0.73
Berthoux et al. [48]ARR score (Berthoux)NoTemporalFranceIgA nephropathy250, 38Death, RRT10, 20No
Bjorneklett [20]Goto scoreYesExternalNorwayIgA nephropathy67633, 146RRT10, 20Coefficients re-estimated, classification simplifiedNo
Magistroni et al. [54]Magistroni scoreNoExternalItalyIgA nephropathy73, 8RRT10No

Hypertensive CKD population.

Population of veterans ≥65 years old. v, variable; (e)GFR, (estimated) glomerular filtration rate in mL/min/1.73 m2; EPV, events per variable/candidate predictor; ‘–’ not reported. The time-frame for which the model performance and other model specifics are reported in bold in the table.

Characteristics of external validation studies and model performance in validations (N = 17) Hypertensive CKD population. Population of veterans ≥65 years old. v, variable; (e)GFR, (estimated) glomerular filtration rate in mL/min/1.73 m2; EPV, events per variable/candidate predictor; ‘–’ not reported. The time-frame for which the model performance and other model specifics are reported in bold in the table.

Risk of bias

Risk of bias was assessed in all 42 included studies, using signalling questions from the PROBAST specified for detecting methodological flaws in both development and validation prediction studies. Overall, the risk of bias was high, as can be seen in Figure 4A and B. Forty-one of 42 studies received a high risk of bias in at least one of the five domains; the only study with an overall low risk of bias was by Schroeder et al. [24]. The majority of studies had a high risk of bias in the domain sample size and missing data. This was often due to the use of complete case analysis, which is generally an inappropriate method of handling missing data. A small sample size was a frequent problem limiting model usage, as a small sample often results in an overfit model and thereby biased results. In the domain statistical analysis, 83% of studies had a high risk of bias. The most common reason was incomplete reporting of performance measures, as few studies reported sufficient calibration results. Also, many studies did not correct their model for overfitting through internal validation. The usability of the model was assessed in a separate domain. If in the publication the full model formula, a calculator or a risk score with absolute risk table was available, then the tool was considered usable. Less than half the studies (48%) presented enough detail for the use of their prediction tool in practice. The usable models that specified a prediction time frame are presented in Figure 5, categorized by the type of patient population and outcome. This figure may be employed as a selection guide when wanting to calculate an individuals’ prognosis, taking into account that many of the models have significant shortcomings and may not be ready for clinical use.
FIGURE 4

(A) Risk of bias and usability of prediction models (N = 42). Assessed using the PROBAST. The five risk of bias domains were evaluated as low risk (+), unclear risk (?) or high risk (−). Usability was evaluated as yes (+) or no (−). (B) PROBAST risk of bias summary for all studies (N = 42).

FIGURE 5

Model selection guide for CKD patients. In this graph, only models that allow calculation of an individual’s prognosis and are therefore labelled as usable are included. This entails that these models provide either a full formula, score with absolute risk table or (currently working) web calculator for a specified prediction time frame. For categories containing multiple models, the risk of bias combined with evidence of external validity was weighed in determining the model order, starting with the most valid and least biased models. Nevertheless, many of the models listed have significant shortcomings and should be used with caution.

(A) Risk of bias and usability of prediction models (N = 42). Assessed using the PROBAST. The five risk of bias domains were evaluated as low risk (+), unclear risk (?) or high risk (−). Usability was evaluated as yes (+) or no (−). (B) PROBAST risk of bias summary for all studies (N = 42). Model selection guide for CKD patients. In this graph, only models that allow calculation of an individual’s prognosis and are therefore labelled as usable are included. This entails that these models provide either a full formula, score with absolute risk table or (currently working) web calculator for a specified prediction time frame. For categories containing multiple models, the risk of bias combined with evidence of external validity was weighed in determining the model order, starting with the most valid and least biased models. Nevertheless, many of the models listed have significant shortcomings and should be used with caution.

DISCUSSION

This systematic review provides an overview of all development and validation studies of predictive models for progression of CKD to kidney failure. Since the last reviews on this topic, the number of publications has more than doubled [3]. Most included studies report high model performance measures, implying that calculating an individual’s risk of renal failure with high accuracy is attainable. This is further emphasized by the similar predictors included in various models. There were, however, substantial shortcomings in many publications. As in many medical prediction studies, aetiological and prediction goals were often confused, limiting interpretability and applicability [7, 58]. First, more than half the tools provided insufficient details to calculate an individual’s prognosis of kidney failure, rendering it useless for its intended purpose. Second, the clinical relevance of many models is limited due to the selection of the derivation population. Third, a high risk of bias was observed across studies, mainly due to the high risk of overfitting, inadequate handling of missing data and incomplete reporting of performance measures. Fourth, sufficient validation was largely lacking, increasing research waste and limiting the reliability of models. And finally, not a single impact study on the effect of clinical uptake has been performed. It is therefore not surprising that clinical uptake of models remains sporadic and guidelines on which model to use are lacking. Providing absolute evidence for the single ‘best’ prognostic tool to use is complicated by differences between studies, mainly concerning varying study populations, use of different prediction baselines, use of varying time frames and multiple outcome definitions. A selection guide including all usable models is presented that may assist clinicians and patients in choosing the tool appropriate to their setting (Figure 5). There are many factors to take into account when selecting the most appropriate model, depending on the user’s wishes and specific clinical setting. Users should be wary of overfitting in models developed on a small sample size and we would advise against the use of these models unless validated in a sufficiently large sample. Based on our results, we would advise the use of a tool with an overall low risk of bias that has shown good performance in external validation in a similar population to the population in which the use is intended and has ideally been assessed in an impact study. For kidney failure prediction in a general CKD cohort with Stages 3–5 patients, we would recommend the four- or eight-variable KFRE, as it has been externally validated extensively for a time frame of 2 and 5 years. Although the development study potentially introduced bias by selecting predictors that were recorded up to 365 days after prediction baseline and by using univariate analysis to select predictors, the model has shown consistently good performance in CKD Stages 3–5 patients from less-biased external validation studies [18, 34]. Alternatively, for 5-year predictions, the Kaiser Permanente Northwest (KPNW) model as updated and externally validated by Schroeder et al. [24] also has great potential, mainly due to its methodological rigor and low risk of bias, although it is less easy to use than the KFRE. Various other general CKD models showed promising results in development but should be further externally validated to ensure consistency of performance before clinical use [26, 28, 32]. For prediction of disease progression in IgA nephropathy patients, a large number of models are available. However, these models, were generally developed on a small sample size and often had a high risk of bias. The most evidence on validity was found for the risk scores developed by Goto et al. [51] and the ARR (by Berthoux et al. [48]). The Goto score contains some risk of bias due to a complete case analysis and univariate selection of predictors, but was developed on a relatively large sample size and has been externally validated twice. Although the ARR score was developed using questionable model building methods and with incomplete reporting of performance, this score has been externally validated the most times, and a recently updated version presented by Knoop et al. [21] shows great potential. Clinical relevance proved to be largely lacking for many of the included models in the current review. Specifically, models for general CKD patients were often developed on prevalent patients with a wide range of disease severity and did not specify a specific time point when the model should be used. Prediction of renal failure can be extremely accurate when using a population with GFRs ranging from 10 to 60 mL/min/1.73 m2. However, in practice, such tools would probably be employed for a more homogeneous group of patients in which it is clinically relevant to discuss prognosis. The predictive capacities of the model would be lower in such a population. This is exemplified in the KFRE validation performed by Peeters et al. [17], where the area under the curve of the four-variable KFRE dramatically decreased from 0.88 in the whole population (CKD Stages 3–5) to 0.71 in the more relevant population of CKD Stage 4 patients. Another factor limiting usability and interpretability is that the number of studies did not define the prediction time frame. Finally, the definition of outcome differs between studies. The use of composite endpoints is particularly problematic, as it limits the value of the model for clinicians, as each separate endpoint requires different interventions. In conclusion, an ideal model is developed for one clearly defined clinically meaningful and objective endpoint in a population for which prediction is clinically relevant. Few models included in this review met these recommendations and this lack of clinical relevance could be a large contributor to the slow uptake seen in practice. Despite the limited uptake and discussed shortcomings of existing tools, risk prediction models for kidney failure have a great potential for improving patients’ decision making, treatment and overall health. In future studies there is the need for improvement of the quality of reporting and methodology used. As the majority of models included had a high risk of bias, these models should not be implemented unless their validity is proven in unbiased external validation studies. Hopefully efforts such as the TRIPOD guidelines will correct these inadequacies and result in more robust, usable and unbiased prognostic tools [9]. To limit research waste and improve clinical uptake, it is of crucial importance that development studies provide enough model information (formula/score with absolute risk table) to enable their use. For specific renal diseases and homogeneous patient populations, there certainly appears to be room for improvement in model development. For populations in which multiple models are available, we advise that future research should focus on the updating, validation and implementation of these existing prognostic tools. Previous studies have shown that the combination of well-established clinical risk factors and kidney disease markers can accurately predict renal failure in a general CKD population. Therefore one might advise focusing resources on updating models for more clinically relevant populations in an unbiased fashion. In this step, external validation of multiple models in the same population is of key importance. Additionally, translation of mathematical model formulas to simple tools such as web calculators and enabling automated uptake is of great importance for integration into daily clinical routine. Ultimately, impact studies will be necessary to determine whether the implementation of such tools truly improves patient outcomes. Ideally, such impact studies would be randomized controlled trials and would assess the effect of implementing a prediction model in clinical practice. Different outcomes might be considered as endpoints in such studies, partly dependent on the time of prediction. Relevant outcomes might be timely referral to nephrologists, timely placement of vascular access, better informed patients, improved quality of life and possibly even improved survival. The current review has a number of strengths. First, we expect to have included a complete overview of existing models. Furthermore, this is the first study on kidney failure models to perform a formal risk of bias assessment aimed specifically at prediction research. The study is limited by the inclusion of only English-language articles. Also, the differences in case mix and characteristics of included studies make it difficult to directly compare their performances. Herein we are limited by the lack of validation studies that compare multiple models in the same cohort. Finally, we limited the scope of this review to models predicting kidney failure, although other outcomes such as death or cardiovascular events may also have significant clinical value. In conclusion, this study provides a systematic overview of existing models for predicting progression to kidney failure in CKD patients. The results may be used as a tool to select the most appropriate and robust prognostic model for various settings. Finally, we hope the current review motivates researchers in this field to decrease the generation of new models and combine efforts to explore, analyse and update existing models in clinically relevant settings to ultimately stimulate clinical uptake and improve patient outcomes.

SUPPLEMENTARY DATA

Supplementary data are available at ndt online.

FUNDING

The work on this study by M.D. and Y.J. was supported by a grant from the Dutch Kidney Foundation (16OKG12). Patient representatives were involved in framing the research question and gave input on the clinical relevance of this project. The Dutch Kidney Foundation was not involved in the study design, interpretation of results or publication approval.

AUTHORS’ CONTRIBUTIONS

All authors have made substantial contributions to the conception of the work and the acquisition and interpretation of data. C.L.R. drafted the work and Y.J., F.W.D. and M.D. critically revised the work. The final version of this manuscript was approved by all the authors and all authors agree to be accountable for all aspects of the work.

CONFLICT OF INTEREST STATEMENT

The authors have no competing interests to declare. All authors declare no support from any organization for the submitted work, no financial relationships with any organizations that might have an interest in the submitted work in the previous 3 years and no other relationships or activities that could appear to have influenced the submitted work. Click here for additional data file.
  57 in total

1.  Prediction versus aetiology: common pitfalls and how to avoid them.

Authors:  Merel van Diepen; Chava L Ramspek; Kitty J Jager; Carmine Zoccali; Friedo W Dekker
Journal:  Nephrol Dial Transplant       Date:  2017-04-01       Impact factor: 5.992

2.  Discussions of the kidney disease trajectory by elderly patients and nephrologists: a qualitative study.

Authors:  Jane O Schell; Uptal D Patel; Karen E Steinhauser; Natalie Ammarell; James A Tulsky
Journal:  Am J Kidney Dis       Date:  2012-01-04       Impact factor: 8.860

3.  Urine biomarkers of tubular injury do not improve on the clinical model predicting chronic kidney disease progression.

Authors:  Chi-Yuan Hsu; Dawei Xie; Sushrut S Waikar; Joseph V Bonventre; Xiaoming Zhang; Venkata Sabbisetti; Theodore E Mifflin; Josef Coresh; Clarissa J Diamantidis; Jiang He; Claudia M Lora; Edgar R Miller; Robert G Nelson; Akinlolu O Ojo; Mahboob Rahman; Jeffrey R Schelling; Francis P Wilson; Paul L Kimmel; Harold I Feldman; Ramachandran S Vasan; Kathleen D Liu
Journal:  Kidney Int       Date:  2016-10-28       Impact factor: 10.612

Review 4.  Risk prediction models for patients with chronic kidney disease: a systematic review.

Authors:  Navdeep Tangri; Georgios D Kitsios; Lesley Ann Inker; John Griffith; David M Naimark; Simon Walker; Claudio Rigatto; Katrin Uhlig; David M Kent; Andrew S Levey
Journal:  Ann Intern Med       Date:  2013-04-16       Impact factor: 25.391

5.  Prediction of ESRD and death among people with CKD: the Chronic Renal Impairment in Birmingham (CRIB) prospective cohort study.

Authors:  Martin J Landray; Jonathan R Emberson; Lisa Blackwell; Tanaji Dasgupta; Rosita Zakeri; Matthew D Morgan; Charlie J Ferro; Susan Vickery; Puja Ayrton; Devaki Nair; R Neil Dalton; Edmund J Lamb; Colin Baigent; Jonathan N Townend; David C Wheeler
Journal:  Am J Kidney Dis       Date:  2010-10-30       Impact factor: 8.860

6.  Biomarkers of inflammation, fibrosis, cardiac stretch and injury predict death but not renal replacement therapy at 1 year in a Canadian chronic kidney disease cohort.

Authors:  Adeera Levin; Claudio Rigatto; Brendan Barrett; François Madore; Norman Muirhead; Daniel Holmes; Catherine M Clase; Mila Tang; Ognjenka Djurdjev
Journal:  Nephrol Dial Transplant       Date:  2013-12-26       Impact factor: 5.992

7.  Looking to the future: predicting renal replacement outcomes in a large community cohort with chronic kidney disease.

Authors:  Angharad Marks; Nicholas Fluck; Gordon J Prescott; Lynn Robertson; William G Simpson; William Cairns Smith; Corri Black
Journal:  Nephrol Dial Transplant       Date:  2015-05-05       Impact factor: 5.992

8.  Risk prediction to inform surveillance of chronic kidney disease in the US Healthcare Safety Net: a cohort study.

Authors:  Yuxiang Xie; Marlena Maziarz; Delphine S Tuot; Glenn M Chertow; Jonathan Himmelfarb; Yoshio N Hall
Journal:  BMC Nephrol       Date:  2016-06-08       Impact factor: 2.388

Review 9.  A systematic review finds prediction models for chronic kidney disease were poorly reported and often developed using inappropriate methods.

Authors:  Gary S Collins; Omar Omar; Milensu Shanyinde; Ly-Mee Yu
Journal:  J Clin Epidemiol       Date:  2012-10-30       Impact factor: 6.437

10.  Validation of the absolute renal risk of dialysis/death in adults with IgA nephropathy secondary to Henoch-Schönlein purpura: a monocentric cohort study.

Authors:  Hesham Mohey; Blandine Laurent; Christophe Mariat; Francois Berthoux
Journal:  BMC Nephrol       Date:  2013-08-01       Impact factor: 2.388

View more
  18 in total

1.  Shared decision-making in advanced kidney disease: a scoping review.

Authors:  Noel Engels; Gretchen N de Graav; Paul van der Nat; Marinus van den Dorpel; Anne M Stiggelbout; Willem Jan Bos
Journal:  BMJ Open       Date:  2022-09-21       Impact factor: 3.006

2.  A prediction model with lifestyle factors improves the predictive ability for renal replacement therapy: a cohort of 442 714 Asian adults.

Authors:  Min-Kuang Tsai; Wayne Gao; Kuo-Liong Chien; Chih-Cheng Hsu; Chi-Pang Wen
Journal:  Clin Kidney J       Date:  2022-04-30

3.  Kidney Failure Prediction Models: A Comprehensive External Validation Study in Patients with Advanced CKD.

Authors:  Chava L Ramspek; Marie Evans; Christoph Wanner; Christiane Drechsler; Nicholas C Chesnaye; Maciej Szymczak; Magdalena Krajewska; Claudia Torino; Gaetana Porto; Samantha Hayward; Fergus Caskey; Friedo W Dekker; Kitty J Jager; Merel van Diepen
Journal:  J Am Soc Nephrol       Date:  2021-03-08       Impact factor: 10.121

4.  Lessons learnt when accounting for competing events in the external validation of time-to-event prognostic models.

Authors:  Chava L Ramspek; Lucy Teece; Kym I E Snell; Marie Evans; Richard D Riley; Maarten van Smeden; Nan van Geloven; Merel van Diepen
Journal:  Int J Epidemiol       Date:  2022-05-09       Impact factor: 9.685

5.  Validation of risk scores for ischaemic stroke in atrial fibrillation across the spectrum of kidney function.

Authors:  Ype de Jong; Edouard L Fu; Merel van Diepen; Marco Trevisan; Karolina Szummer; Friedo W Dekker; Juan J Carrero; Gurbey Ocak
Journal:  Eur Heart J       Date:  2021-04-14       Impact factor: 29.983

6.  Impact of Using Risk-Based Stratification on Referral of Patients With Chronic Kidney Disease From Primary Care to Specialist Care in the United Kingdom.

Authors:  Harjeet K Bhachu; Paul Cockwell; Anuradhaa Subramanian; Nicola J Adderley; Krishna Gokhale; Anthony Fenton; Derek Kyte; Krishnarajah Nirantharakumar; Melanie Calvert
Journal:  Kidney Int Rep       Date:  2021-06-01

7.  Person centred care provision and care planning in chronic kidney disease: which outcomes matter? A systematic review and thematic synthesis of qualitative studies : Care planning in CKD: which outcomes matter?

Authors:  Ype de Jong; Esmee M van der Willik; Jet Milders; Yvette Meuleman; Rachael L Morton; Friedo W Dekker; Merel van Diepen
Journal:  BMC Nephrol       Date:  2021-09-13       Impact factor: 2.388

8.  Associations between urinary cysteine-rich protein 61 excretion and kidney function decline in outpatients with chronic kidney disease: a prospective cohort study in Taiwan.

Authors:  Chun-Fu Lai; Jian-Jhong Wang; Ya-Chun Tu; Chia-Yu Hsu; Hon-Yen Wu; Cheng-Chung Fang; Yung-Ming Chen; Ming-Shiou Wu; Tun-Jun Tsai
Journal:  BMJ Open       Date:  2021-10-06       Impact factor: 2.692

9.  CKD subpopulations defined by risk-factors: A longitudinal analysis of electronic health records.

Authors:  Rajagopalan Ramaswamy; Soon Nan Wee; Kavya George; Abhijit Ghosh; Joydeep Sarkar; Rolf Burghaus; Jörg Lippert
Journal:  CPT Pharmacometrics Syst Pharmacol       Date:  2021-09-12

10.  Use of the kidney failure risk equation to inform clinical care of patients with chronic kidney disease: a mixed-methods systematic review.

Authors:  Harjeet Kaur Bhachu; Anthony Fenton; Paul Cockwell; Olalekan Aiyegbusi; Derek Kyte; Melanie Calvert
Journal:  BMJ Open       Date:  2022-01-18       Impact factor: 2.692

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.