| Literature DB >> 31368172 |
Rachel Heyard1, Jean-François Timsit2, Leonhard Held1.
Abstract
Clinical prediction models play a key role in risk stratification, therapy assignment and many other fields of medical decision making. Before they can enter clinical practice, their usefulness has to be demonstrated using systematic validation. Methods to assess their predictive performance have been proposed for continuous, binary, and time-to-event outcomes, but the literature on validation methods for discrete time-to-event models with competing risks is sparse. The present paper tries to fill this gap and proposes new methodology to quantify discrimination, calibration, and prediction error (PE) for discrete time-to-event outcomes in the presence of competing risks. In our case study, the goal was to predict the risk of ventilator-associated pneumonia (VAP) attributed to Pseudomonas aeruginosa in intensive care units (ICUs). Competing events are extubation, death, and VAP due to other bacteria. The aim of this application is to validate complex prediction models developed in previous work on more recently available validation data.Entities:
Keywords: area under the curve; calibration slope; competing events; discrete time-to-event model; dynamic prediction models; prediction error; validation
Mesh:
Year: 2019 PMID: 31368172 PMCID: PMC7217187 DOI: 10.1002/bimj.201800293
Source DB: PubMed Journal: Biom J ISSN: 0323-3847 Impact factor: 2.207
Total number of distinct events in the training and testing data
| Dead | Extubated | VAP noPA | VAP PA | |
|---|---|---|---|---|
| Training data | 896 | 3,251 | 635 | 341 |
| Testing data | 126 | 312 | 72 | 24 |
Notes:. The information of the patients is only analyzed until the occurrence of a first event.
The number of distinct daily events in the testing data set during the first 3 weeks of ventilation
| Days since start of ventilation ( | ||||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Event cause | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16 | 17 | 18 | 19 | 20 | 21 |
|
| Dead | 22 | 20 | 5 | 13 | 9 | 7 | 12 | 5 | 3 | 5 | 2 | 0 | 4 | 2 | 2 | 3 | 1 | 115 |
| Extubated | 38 | 52 | 44 | 35 | 25 | 18 | 13 | 15 | 13 | 6 | 3 | 5 | 8 | 4 | 2 | 5 | 1 | 287 |
| VAP noPA | 17 | 7 | 11 | 5 | 7 | 4 | 3 | 3 | 3 | 2 | 3 | 3 | 0 | 0 | 0 | 1 | 1 | 70 |
| VAP PA | 3 | 3 | 6 | 0 | 1 | 1 | 3 | 0 | 1 | 0 | 3 | 0 | 0 | 0 | 0 | 0 | 1 | 22 |
Figure 1Cause‐specific AUC curves for the three models of interest following Li et al. (2018)
The cause‐specific C‐indices with their 95% jackknife confidence intervals, for the three prediction models
| Model | Model | Model | ||||
|---|---|---|---|---|---|---|
| Dead | 0.751 | [0.70, 0.80] | 0.749 | [0.70, 0.80] | 0.748 | [0.69, 0.80] |
| Extubated | 0.689 | [0.65, 0.72] | 0.700 | [0.67, 0.73] | 0.701 | [0.67, 0.73] |
| VAP noPA | 0.684 | [0.63, 0.74] | 0.689 | [0.63, 0.75] | 0.658 | [0.60, 0.72] |
| VAP PA | 0.583 | [0.41, 0.76] | 0.585 | [0.41, 0.76] | 0.575 | [0.41, 0.74] |
The overall cause‐independent C‐index (7) with 95% jackknife confidence interval for all models
| Model | Model | Model | |||
|---|---|---|---|---|---|
| 0.698 | [0.67, 0.72] | 0.705 | [0.68, 0.73] | 0.700 | [0.67, 0.73] |
Figure 2Cause‐specific calibration plots for the different models following Berger and Schmid. The dashed line is the ideal 45 degree line indicating, while the solid lines are simple regression lines
Cause‐specific calibration slopes and intercepts for the three prediction models with their 95% Wald confidence intervals and the p‐values of the joint (test 1) and separate (test 2) likelihood ratio tests as discussed in Section 3.2
| Model | Model | Model | ||||
|---|---|---|---|---|---|---|
| CS | CI | CS | CI | CS | CI | |
| Dead | 0.97 [0.79; 1.16] | 0.39 [0.20; 0.58] | 0.98 [0.79; 1.16] | 0.45 [0.26; 0.64] | 0.98 [0.80; 1.16] | 0.45 [0.25; 0.64] |
| Extubated | 0.90 [0.71; 1.09] | 0.15 [0.03; 0.27] | 0.83 [0.65; 1.01] | 0.35 [0.22; 0.47] | 0.82 [0.65; 1.00] | 0.35 [0.23; 0.47] |
| VAP noPA | 1.12 [0.69; 1.55] | 0.55 [0.31; 0.79] | 0.94 [0.55; 1.33] | 0.73 [0.49; 0.96] | 0.94 [0.54; 1.33] | 0.69 [0.45; 0.93] |
| VAP PA | 0.69 [0.02; 1.36] | 0.07 [−0.35; 0.49] | 0.55 [−0.05; 1.14] | 0.07 [−0.35; 0.49] | 0.69 [0.03; 1.34] | 0.04 [−0.38; 0.46] |
|
| <.0001 | <.0001 | < 0.0001 | |||
|
| .69 | <.0001 | .23 | <.0001 | .32 | <.0001 |
Figure 3Cause‐specific prediction error curves for the three models of interest, with the number of daily events
Figure 4Cause‐specific relative error reduction curves for different prediction models
The cause‐specific integrated prediction errors (× 100) together with their 95% jackknife confidence intervals for the different prediction models
| Model | Model | Model | Model | |||||
|---|---|---|---|---|---|---|---|---|
| Dead | 2.188 | [1.51; 2.86] | 1.815 | [1.25; 2.38] | 1.807 | [1.25; 2.37] | 1.813 | [1.25; 2.38] |
| Extubated | 10.134 | [9.02; 11.25] | 8.932 | [7.88; 9.98] | 8.837 | [7.75; 9.92] | 8.832 | [7.75; 9.92] |
| VAP noPA | 0.877 | [0.50; 1.25] | 0.853 | [0.49; 1.21] | 0.852 | [0.49; 1.21] | 0.850 | [0.49; 1.21] |
| VAP PA | 0.089 | [0.019; 0.159] | 0.088 | [0.018; 0.157] | 0.088 | [0.019; 0.157] | 0.088 | [0.019; 0.156] |