| Literature DB >> 34138495 |
Ype de Jong1,2, Chava L Ramspek1, Carmine Zoccali3,4, Kitty J Jager5, Friedo W Dekker1, Merel van Diepen1.
Abstract
Over the past few years, a large number of prediction models have been published, often of poor methodological quality. Seemingly objective and straightforward, prediction models provide a risk estimate for the outcome of interest, usually based on readily available clinical information. Yet, using models of substandard methodological rigour, especially without external validation, may result in incorrect risk estimates and consequently misclassification. To assess and combat bias in prediction research the prediction model risk of bias assessment tool (PROBAST) was published in 2019. This risk of bias (ROB) tool includes four domains and 20 signalling questions highlighting methodological flaws, and provides guidance in assessing the applicability of the model. In this paper, the PROBAST will be discussed, along with an in-depth review of two commonly encountered pitfalls in prediction modelling that may induce bias: overfitting and composite endpoints. We illustrate the prevalence of potential bias in prediction models with a meta-review of 50 systematic reviews that used the PROBAST to appraise their included studies, thus including 1510 different studies on 2104 prediction models. All domains showed an unclear or high ROB; these results were markedly stable over time, highlighting the urgent need for attention on bias in prediction research. This article aims to do just that by providing (1) the clinician with tools to evaluate the (methodological) quality of a clinical prediction model, (2) the researcher working on a review with methods to appraise the included models, and (3) the researcher developing a model with suggestions to improve model quality.Entities:
Keywords: clinical epidemiology; epidemiology; evidence-based medicine; medical education; meta-analysis
Mesh:
Year: 2021 PMID: 34138495 PMCID: PMC9291738 DOI: 10.1111/nep.13913
Source DB: PubMed Journal: Nephrology (Carlton) ISSN: 1320-5358 Impact factor: 2.358
FIGURE 1The increase in the number of prediction studies in PubMed (for the search string, see Data S1)
The domains and signalling questions of the PROBAST for assessment of risk of bias and applicability. Data presented with permission of Wolff, coauthor of the PROBAST
| Signalling questions | ||||
|---|---|---|---|---|
| 1. Participants | 2. Predictors | 3. Outcome | 4. Analysis | |
| Risk of bias | 1.1. Were appropriate data sources used, for example, cohort, RCT, or nested case–control study data? | 2.1. Were predictors defined and assessed in a similar way for all participants? | 3.1. Was the outcome determined appropriately? | 4.1. Were there a reasonable number of participants with the outcome? |
| 1.2. Were all inclusions and exclusions of participants appropriate? | 2.2. Were predictor assessments made without knowledge of outcome data? | 3.2. Was a prespecified or standard outcome definition used? | 4.2. Were continuous and categorical predictors handled appropriately? | |
| ‐ | 2.3. Are all predictors available at the time the model is intended to be used? | 3.3. Were predictors excluded from the outcome definition? | 4.3. Were all enrolled participants included in the analysis? | |
| ‐ | ‐ | 3.4. Was the outcome defined and determined in a similar way for all participants? | 4.4. Were participants with missing data handled appropriately? | |
| ‐ | ‐ | 3.5. Was the outcome determined without knowledge of predictor information? | 4.5. Was selection of predictors based on univariable analysis avoided? | |
| ‐ | ‐ | 3.6. Was the time interval between predictor assessment and outcome determination appropriate? | 4.6. Were complexities in the data (e.g., censoring, competing risks, sampling of control participants) accounted for appropriately? | |
| ‐ | ‐ | ‐ | 4.7. Were relevant model performance measures evaluated appropriately? | |
| ‐ | ‐ | ‐ | 4.8. Were model overfitting, underfitting, and optimism in model performance accounted for? | |
| ‐ | ‐ | ‐ | 4.9. Do predictors and their assigned weights in the final model correspond to the results from the reported multivariable analysis? | |
FIGURE 2Model fitting illustrated. Different types of fit in candidate predictor selection, illustrated by two hypothetical samples of n = 30: a development cohort on the left, and a validation cohort on the right. Dots indicate the outcome risk for the predictor value (black dots in the development cohort; red dots in the validation cohort); the blue line indicates the fitted model. BMI; Body Mass Index
FIGURE 3Aggregated overview of the risk of bias in 1039 prediction models with complete data (as assessed with the PROBAST in 50 systematic reviews)
FIGURE 4R Risk of bias of 1039 prediction models extracted from 50 systematic reviews with complete data as assessed with the PROBAST, stratified per year of publication and domain. Nine hundred and eighty five models presented information on the bias domains, and 560 presented information on the applicability domains. The trend is indicated by a fitted LOESS trendline with 95% confidence interval. For clarity, data points are jittered on the y‐axis, by adding a Gaussian error with a standard deviation of 0.1