| Literature DB >> 31093561 |
Teus H Kappen1, Wilton A van Klei1, Leo van Wolfswinkel1, Cor J Kalkman1, Yvonne Vergouwe2,3, Karel G M Moons1,3,4.
Abstract
An important aim of clinical prediction models is to positively impact clinical decision making and subsequent patient outcomes. The impact on clinical decision making and patient outcome can be quantified in prospective comparative-ideally cluster-randomized-studies, known as 'impact studies'. However, such impact studies often require a lot of time and resources, especially when they are (cluster-)randomized studies. Before envisioning such large-scale randomized impact study, it is important to ensure a reasonable chance that the use of the prediction model by the targeted healthcare professionals and patients will indeed have a positive effect on both decision making and subsequent outcomes. We recently performed two differently designed, prospective impact studies on a clinical prediction model to be used in surgical patients. Both studies taught us new valuable lessons on several aspects of prediction model impact studies, and which considerations may guide researchers in their decision to conduct a prospective comparative impact study. We provide considerations on how to prepare a prediction model for implementation in practice, how to present the model predictions, and how to choose the proper design for a prediction model impact study.Entities:
Keywords: Diagnosis; Impact studies; Implementation; Prediction models; Prognosis; Study design
Year: 2018 PMID: 31093561 PMCID: PMC6460651 DOI: 10.1186/s41512-018-0033-6
Source DB: PubMed Journal: Diagn Progn Res ISSN: 2397-7523
Fig. 1Methodological similarities and differences between our cluster-randomized trial with an assistive prediction tool (first study) and the before-after study with a directive prediction tool (second study)
Predictors and regression coefficients of the original and updated models
| Original model | Updated model | ||
|---|---|---|---|
| Predictor | Regression coefficients | Predictor | |
| Age (years) | − 0.022 | − 0.017 | Age (years) |
| Female gender | 0.46 | 0.36 | Female gender |
| Current smoking | − 0.63 | − 0.50 | Current smoking |
| History of PONV or motion sickness | 0.76 | 0.60 | History of PONV or motion sickness |
| Lower abdominal or middle ear surgery | 0.61 | 0.48 | Abdominal or middle ear surgerya |
| Isoflurane and/or nitrous oxide anesthesiab | 0.72 | 0.35 | Inhalational anesthesiab |
| – | − 1.16 | Outpatient surgeryc | |
| Intercept | 0.15 | 0.12 | Intercept |
| Model performance characteristicsd | |||
| Model discrimination as C-statistic (95% CI) | 0.62 (0.60–0.64) | 0.68 (0.66–0.70) | Model discrimination as C-statistic (95% CI) |
| Calibration slope (95% CI) | 0.57 (0.48–0.66) | 1.00 (0.89–1.10) | Calibration slope (95% CI) |
PONV postoperative nausea and vomiting, CI confidence interval
aIn the updated model, the predictor included lower abdominal, upper abdominal, and laparoscopic surgery in addition to middle ear surgery
bAs compared to intravenous anesthesia using propofol
cPredictor not included in the original model, but added in the update of the model
dModel performance was validated in a subset of patients (between March 2006 and February 2007) treated by anesthesiologists of the care-as-usual group of the cluster-randomized trial [19]
Is the prediction model ready for implementation?
| 1. Assess the current state of scientific evidence | |
| A prediction model should at least have been validated once to assess its predictive performance in new patients or in a new setting. Subsequent diagnostic and therapeutic steps should also have a valid scientific base. | |
| 2. Verify the predictive performance of the prediction model in the new setting | |
| Local practice, medical care, and patient population may not be similar to the setting in which the prediction model was derived. Consider possible differences between the two settings. | |
| 3. Tailor the prediction model to optimize the predictive performance in the new setting | |
| An insufficient predictive performance in the new setting requires a model update. Even simple adjustments may overcome poor performance in the new setting. | |
| 4. Develop a real-time strategy to handle missing predictor values when using the model | |
| Multivariable imputation is preferred over simply omitting predictors. Other predictors of the model, additional patient information, and information about the local clinical process may be used to estimate missing predictor values. |
How to present the model predictions?
| Facilitators: features that increase the ease of use of a prediction model | |
| F.1 Add a decision recommendation to the predicted probabilities | |
| Directive prediction tools may be easier for physicians to use in their decision making than assistive prediction tools that provide only predicted probabilities without decision recommendations. | |
| F.2 Automatic calculation and presentation of the model’s probability within the physician’s workflow | |
| Minimizing manual predictor value entry and integrating the estimation of the model’s probability in the electronic patient record will facilitate the ease of use of a prediction model for care providers. | |
| F.3 Provide the reasoning or research evidence behind the predicted probability | |
| Enhances face value, acceptation and belief in the model, and thus the willingness to use the model’s probabilities to guide decision making. | |
| Barriers that may decrease the ease of use of a prediction model | |
| B.1 A predicted probability may be difficult to use in decision making, especially without corresponding recommendations | |
| Weighing the numerical probabilities with other available information will require more cognitive effort from physicians when the probabilities are presented without a corresponding recommendation on subsequent treatment or additional diagnostic testing. | |
| B.2 When the targeted physicians use an intuitive rather than analytical process of decision making | |
| When an existing decision-making process is mostly intuitive, it may require more cognitive effort to use probabilistic knowledge in decision making. | |
| B.3 When the predicted outcome is not a main concern for the physicians | |
| Physicians will not prioritize their time and efforts to use a prediction model in their decisions, when they consider other problems or outcomes to be more important. | |
| B.4 A prediction model does not weigh the benefits and risks of treatment or additional diagnostics regarding the patient’s (co)morbidity | |
| When a physician has more sources of information about the benefits and risks of subsequent treatment decisions, she/he will still have to weigh the model’s predicted probability from the model against this information, which is often perceived as cumbersome. |
Design of the impact study
| 1. Consider a decision analytic study | |
| If not previously performed, a decision analytic study may link the available evidence to estimate the theoretical impact on decision making and/or patient outcome. | |
| 2. Consider studying the effects on both physician behavior and patient outcome | |
| Changes in process or behavior may not be sufficient to improve patient outcome. Studying the effects on patient outcome typically requires more time and money. | |
| 3. Consider additional data collection to improve the understanding of the impact study results | |
| The impact does not only depend on the prediction model, but also on physician decision making and the effectiveness of subsequent treatment. Without additional data that is collected during the impact study, the effects of the individual components may difficult to disentangle. | |
| 4. Compare the use of a prediction model to care-as-usual | |
| Physicians are not naive in patient selection and making interventional decisions. The impact of a prediction model (assistive or directive) is its value over and above current clinical decision making. | |
| 5. Cluster-randomized trial as the optimal design | |
| Randomization of practices or practitioners aims to prevent learning effects and contamination between study groups. Nonetheless, time and costs to perform a cluster-randomized study should be weighed against its expected informational value. | |
| 6. Consider using each study group as its own control | |
| The balance between study groups may be improved when using a stepped wedge design and including pre-trial observations. | |
| 7. The impact of the prediction model will depend on the predicted probabilitya | |
| Predicted probability should be considered an effect modifier in the statistical analysis, which requires, e.g. stratification or use of its interaction term with ‘study group’ in regression analyses. | |
| 8. All predictors should be available for care-as-usual patientsa | |
| A probability-dependent analysis of the results requires that the predicted probabilities can afterwards also be estimated for the care-as-usual patients (control group). Accordingly, all predictors must be available for care-as-usual patients, even the costly or invasive predictor variables. |
aAdditional item, not further explained in the manuscript
Fig. 2Graphical representation of the mixed effects regression analysis on the pooled dataset of the two impact studies. In both studies, a prediction model for postoperative nausea and vomiting (PONV) was implemented (red colors) and compared to care-as-usual (blue colors). In one study, the probabilities of the model were simply presented to physicians who implemented without a recommendation (i.e. an assistive format; less saturated colors). The other study also included an actionable recommendation (i.e. a directive format; more saturated colors). Both studies compared the impact on the physicians’ administration of antiemetic prophylaxis (a) and on the incidence of PONV (b). The bars and their 95% confidence intervals (CI) represent the fixed effects of the mixed effects regression analyses. The mixed effects models included fixed effects for the following variables: study, allocation group, predicted probability of PONV, and all interaction term between these variables. Because of the similarity of the results, the bars were calculated from the unadjusted analysis after multiple imputations. The 95% CIs were calculated from the covariance matrix for the variable study, allocation group, predicted probability, and their interaction terms. Further methodologic information and the numerical results of the regression models are available in Additional file 1