| Literature DB >> 33564405 |
Chava L Ramspek1, Kitty J Jager2, Friedo W Dekker1, Carmine Zoccali3, Merel van Diepen1.
Abstract
Prognostic models that aim to improve the prediction of clinical events, individualized treatment and decision-making are increasingly being developed and published. However, relatively few models are externally validated and validation by independent researchers is rare. External validation is necessary to determine a prediction model's reproducibility and generalizability to new and different patients. Various methodological considerations are important when assessing or designing an external validation study. In this article, an overview is provided of these considerations, starting with what external validation is, what types of external validation can be distinguished and why such studies are a crucial step towards the clinical implementation of accurate prediction models. Statistical analyses and interpretation of external validation results are reviewed in an intuitive manner and considerations for selecting an appropriate existing prediction model and external validation population are discussed. This study enables clinicians and researchers to gain a deeper understanding of how to interpret model validation results and how to translate these results to their own patient population.Entities:
Keywords: educational; external validation; methodology; prediction models
Year: 2020 PMID: 33564405 PMCID: PMC7857818 DOI: 10.1093/ckj/sfaa188
Source DB: PubMed Journal: Clin Kidney J ISSN: 2048-8505
FIGURE 1Illustration of different validation types. A developed prediction model can be validated in various ways and in populations that differ from the development cohort to varying degrees. Internal validation uses the patients from the development population and can therefore always be performed. As internal validation does not include new patients, it mainly provides information on the reproducibility of the prediction model. Temporal validation is often considered to lie midway between internal and external validation. It entails validating the model on new patients who were included in the same study as patients from the development cohort but sampled at an earlier or later time point. It provides some information on both the reproducibility and generalizability of a model. External validation mainly provides evidence on the generalizability to various different patient populations. Patients included in external validation studies may differ from the development population in various ways: they may be from different countries (geographic validation), from different types of care facilities or have different general characteristics (e.g. frail older patients versus fit young patients). Not every model needs to be validated in all the ways depicted. In certain cases, internal validation or only geographic external validation may be sufficient; this is dependent on the research question and size of the development cohort.
FIGURE 2Cumulative histogram of the number of hits on PubMed when using a simple search strategy of prediction models and adding external validation to this search. Search strategies are given in Appendix A. PubMed was searched from 1961 to 2019. The total number of prediction model studies retrieved was 84 032, of which 4309 were found when adding an external validation search term. The percentage of studies with external validation increased over the years; in 1990, 0.5% of published prediction studies mentioned external validation, while in 2019 this was 7%.
FIGURE 3Example of a calibration plot. The dotted line at 45 degrees indicates perfect calibration, as predicted and observed probabilities are equal. The 10 dots represent tenths of the population divided based on predicted probability. The 10% of patients with the lowest predicted probability are grouped together. Within this group the average predicted risk and proportion of patients who experience the outcome (observed probability) are computed. This is repeated for subsequent tenths of the patient population. The blue line is a smoothed lowess line. For a logistic model this is computed by plotting each patient individually according to their predicted probability and outcome (0 or 1) and plotting a flexible averaged line based on these points. In this example calibration plot we can see that the model overpredicts risk; when the predicted risk is 60%, the observed risk is ∼35%. This overprediction is more extreme for the high-risk x-axis. If a prediction model has suggested cut-off points for risk groups, then we recommend plotting these various risk groups in the calibration plot (instead of tenths of the population).
Difference in discriminatory performance of mortality prediction models when validated on a population of HD versus PD patients
| Prediction model | Original | Discrimination: C-statistic | |
|---|---|---|---|
| Population | HD | PD | |
| Floege 1 | HD | 0.70 | 0.78 |
| Floege 2 | HD | 0.71 | 0.78 |
| Holme | HD | 0.71 | 0.77 |
| Mauri | HD | 0.67 | 0.80 |
| Hutchinson | HD | 0.67 | 0.77 |
All prediction models listed were exclusively developed on an HD population. This table was adapted from Table 4 published in a study by Ramspek et al., with permission [3].
Key points, dos and don’ts concerning the external validation of prognostic models
|
|
|
|
|---|---|---|
| What is external validation? | ||
|
External validation is testing a prediction model in new individuals External validation cohorts may differ from the development cohort in geographic location, care-setting or patient characteristics | Do externally validate prediction models in separate studies and by independent researchers | Do not perform a random split-sample validation; this is an inefficient type of internal validation |
| Why is external validation important? | ||
|
External validation is needed to determine a model’s reproducibility and transportability Most developed models are never validated or used, which leads to significant research waste. | Do assess a prediction model’s transportability for each population in which clinical usage is desired | Do not implement a prediction model in clinical practice before external validity has been established |
| How does external validation of a prediction model work? | ||
|
Validating a prediction model essentially means comparing predicted risks to observed outcomes Discrimination and calibration are the most important elements of model performance | Do externally validate the model in the form which is intended for use; this may be a simplified risk score | Do not extensively update a prediction model without subsequently determining its external validity in new individuals |
| When is a prediction model suitable for external validation? | ||
|
Prediction models which are appropriate for the intended clinical use, regarding predictors and outcome, are suitable for external validation Models which allow an individual’s absolute risk calculation, were developed with a low risk of bias and show relatively good predictive performance in previous validation are preferred | Do assess whether design flaws in model development cause biased predictions by correcting these flaws in the external validation | Do not externally validate an existing model in the development cohort of a new model; the new model will almost always seem superior |
| Where should a prediction model be externally validated? Choosing the validation cohort | ||
|
The ideal validation population is a large observational cohort which is designed to accurately represent a specific clinical patient population Differences in predictive performance between validation cohorts may be caused by heterogeneity in predictor effects, varying outcome incidence and differences in case-mix | Do report the degree of relatedness between development and external validation cohorts | Do not combine heterogeneous subgroups to assess whether a prediction model works for everybody, as model discrimination will be deceptively good |