Literature DB >> 20846460

The search for stable prognostic models in multiple imputed data sets.

David Vergouw1, Martijn W Heymans, George M Peat, Ton Kuijpers, Peter R Croft, Henrica C W de Vet, Henriëtte E van der Horst, Daniëlle A W M van der Windt.   

Abstract

BACKGROUND: In prognostic studies model instability and missing data can be troubling factors. Proposed methods for handling these situations are bootstrapping (B) and Multiple imputation (MI). The authors examined the influence of these methods on model composition.
METHODS: Models were constructed using a cohort of 587 patients consulting between January 2001 and January 2003 with a shoulder problem in general practice in the Netherlands (the Dutch Shoulder Study). Outcome measures were persistent shoulder disability and persistent shoulder pain. Potential predictors included socio-demographic variables, characteristics of the pain problem, physical activity and psychosocial factors. Model composition and performance (calibration and discrimination) were assessed for models using a complete case analysis, MI, bootstrapping or both MI and bootstrapping.
RESULTS: Results showed that model composition varied between models as a result of how missing data was handled and that bootstrapping provided additional information on the stability of the selected prognostic model.
CONCLUSION: In prognostic modeling missing data needs to be handled by MI and bootstrap model selection is advised in order to provide information on model stability.

Entities:  

Mesh:

Year:  2010        PMID: 20846460      PMCID: PMC2954918          DOI: 10.1186/1471-2288-10-81

Source DB:  PubMed          Journal:  BMC Med Res Methodol        ISSN: 1471-2288            Impact factor:   4.615


Background

In healthcare predicting how long it takes for an episode of musculoskeletal pain to resolve can be difficult. Outcome varies between patients and over time. Although clinicians can be relatively good "prognosticians" [1,2] clinical judgment and intuition can be incorrect and difficult to quantify or to be made explicit. To understand the ingredients that contribute to correct prognosis and to improve upon clinical judgment, clinical prediction rules can be useful. These provide a quantitative estimate of the absolute risk of particular outcomes of interest for individual patients, which may subsequently be used to support decisions regarding treatment. Until now, several clinical prediction rules have been developed in the field of musculoskeletal pain, for example to estimate the outcome of low back [2-4], knee [5] or shoulder pain [6]. In the development of clinical prediction models, researchers frequently use a regression analysis with a backward or forward selection strategy. However, this methodology may result in overoptimistically estimated regression coefficients, omission of important predictors and random selection of less important predictors. As a result derived models may be unstable [7]. Incorporating a bootstrap resampling procedure in model development has been suggested to provide information on model stability [8-11]. Since bootstrapping mimics the sampling variation in the population from which the sample was drawn it is expected to produce a model which better represents the underlying population [9-11]. Another problem occurring in prognostic studies is missing data. Multiple imputation (MI), which uses all observed information, was shown to be superior to other imputation techniques like single regression imputation [12,13]. Though, MI is not yet frequently used in predictive modelling and model stability is hardly ever accounted for in MI approaches. It has been shown that for low back pain extending MI with a bootstrapping procedure provides an accurate model selection and information on model stability [14]. However generalizability of this method was never tested in other patient datasets. Therefore, the objective of our research was to examine the influence of bootstrapping and multiple imputation on model composition and stability in a shoulder pain data set with missing values.

Methods

Study population

We used data from the Dutch Shoulder Study (DSS) [15]. This cohort consists of 587 patients who consulted their general practitioner (GP) with a new episode of shoulder disorders. Inclusion criteria were: no GP consultation or treatment received for the afflicted shoulder in the preceding three months. Exclusion criteria were: dementia, severe psychiatric of physical conditions (i.e. fractures or dislocation in the shoulder region, rheumatic diseases, neoplasms, neurological of vascular disorders). The ethics review board of the VU University medical centre approved the study protocol.

Outcome measures

We focused on two outcome measures; persistent shoulder disability (16-item SDQ; 0-100) [16] and persistent shoulder pain intensity (Numeric Rating Scale; 0-10) [17]. To define 'persistence' baseline scores were subtracted from follow-up scores. An optimal cut-off point was defined by studying the relationship between the change scores and a secondary outcome measure 'patient perceived recovery' [6]. Patients were denoted as recovered when they characterized their complaints as 'fully recovered' or 'very much improved'. By constructing Receiver Operating Characteristic (ROC) curves with patient perceived recovery as the external criterion, the optimal cutoff point (i.e. that point that yields the lowest overall misclassification) was determined [18]. According to this analysis a 50% decrease in disability and pain intensity compared to baseline was considered a minimal important change, and was used as a cut-off value to dichotomize both outcome measures. Patients who improved less then 50% were denoted as having persistent pain or disability. Outcomes were measured three months after enrolment by postal questionnaire.

Prognostic factors

Based on a systematic review of the literature [19] a set of candidate predictors was selected, including demographic variables, characteristics of the shoulder pain problem, physical and psychological factors (see Table 1). The following questionnaires were used to gather information on psychological factors: the Pain Coping and Cognition List (PCCL [20]: pain coping, catastrophizing, internal and external locus of control), the 4 Dimensional Symptom Questionnaire (4DSQ [21]: anxiety, depression, somatisation, distress), the Fear-Avoidance Beliefs Questionnaire (FABQ [22]: fear-avoidance) and the Tampa Scale for Kinesiophobia (TSK [23,24]: kinesiophobia). Within 10 days after consulting the GP, participants completed a baseline questionnaire to assess potential predictors.
Table 1

Patient characteristics at baseline

variablen(%)
Demographic
 Age (years); mean (SD)51(14)
 gender (male)292(50)
 education
  low*210(36)
  middle234(40)
  high135(23)
Disease characteristics
 shoulder complaints in the past year321(55)
 neck complaints in the past year252(43)
 duration of complaints
  0-6 weeks205(35)
  7-12 weeks139(24)
  > 3 months242(41)
 gradual onset (vs. acute)363(62)
 shoulder pain (0-10); mean (SD)4.8(2)
 shoulder disability (0-100); mean (SD)59.9(24)
 both shoulders afflicted74(13)
 co-morbidity469(80)
  upper extremity joint pain245(42)
  neck pain197(34)
  upper extremity joint pain174(30)
  low back pain139(24)
  high back pain53(9)
Psychological factors
 psychological complaints55(9)
 pain coping (0-6); mean (SD)2.98(0.98)
 catastrophizing (0-6); mean (SD)2.2(0.8)
 internal locus of control (0-6); mean (SD)3.3(0.9)
 external locus of control(0-6); mean (SD)3.2(0.88)
 anxiety (0-6); mean (SD)0.3(1.2)
 depression (0-6); mean (SD)0.2(1.3)
 somatisation (0-6); mean (SD)3.3(4.1)
 distress (0-6); mean (SD)2.3(4.5)
 fear-avoidance (0-6); mean (SD)14.1(5.6)
 kinesophobia (0-6); mean (SD)3.2(3.5)
Physical factors
 physical load at work (0-5); mean (SD)1.2(1.5)
 physical activity
  less active than others110(39)
  equally active245(42)
  more active226(19)
 inability to perform daily activities last year
  1-30 days184(31)
  1-12 months61(10)
 sporting activities230(39)
 cause of shoulder problem: sporting injury29(5)

SD = standard deviation

Patient characteristics at baseline SD = standard deviation

Analysis

For all continuous predictors the linearity assumption was checked. When the relationships between variables and outcome did not resemble linearity, variables were categorized (3 categories) or dichotomized. Although this causes loss of information [25], these procedures were retained since they are part of the frequently used standard statistical methodology in predictive modeling. Variables were checked for (multi)collinearity using Pearson's r, given that correlated variables can disturb variable selection in multivariable regression [26]. In case of correlation (r≥0.5) the variables which could most easily be obtained in clinical practice by the physician were retained. To reduce the initial number of variables, an univariable analysis (α > 0.157) was performed in both the imputed and unimputed data sets, thus all analyses were preceded by this pre-selection. The subsequent analyses were all based on a multivariable analysis with a backward selection strategy and a stopping rule of α = 0.157. This significance level is available in many statistical software packages and results have been shown to be comparable with the more complex Akaike Information Criterion (AIC) [27]. The number of events per variable (EPV) was calculated for each method to check whether the analysis was sufficiently powered (EPV > 10) [28]. The checklist proposed by Harrell [29] for multivariable modeling was followed where possible. To study the effect of missing data and model stability on model composition, the following four methods were compared:

1) Complete Case Analysis (CCA)

To handle missing data, subjects with missing values on any of the variables were omitted and only those subjects with information on all variables in the model were included for analysis.

2) Multiple imputation (MI-5)

Missing values were imputed using a Multivariate Imputation by Chained Equations (MICE) procedure with the "predictive mean matching" as imputation method [30]. All available data including outcome measure were used in the imputation method [13]. We generated five imputed data sets (MI-5). Multivariable regression was applied to each of the 5 imputed data sets. From these 5 models, predictors which appeared in at least 2 models (a Inclusion Fraction of ≥40%) qualified for the final model. Whether these predictors significantly contributed to the final model was tested using a likelihood ratio test [31] with a critical P-value of P = 0.157. Predictors were dropped from the final model in case of a nonsignificant (P > 0.157) likelihood ratio.

3) Bootstrapping (B)

A two-step bootstrap model selection procedure [9,11] was applied to provide information on model stability. First 500 samples with replacement were taken from the complete case data set. In each sample a multivariable model was built. To be consistent with the MI-5 method, predictors which appeared in ≥40% of these models qualified for the second step. In this second step 500 new complete case samples were taken and in each of which a multivariable model was built using the predictors from the first step. These 500 models provided information on model stability (i.e. which combination of predictors is most frequently selected in the model).

4) Multiple imputation + bootstrapping (MI-5+B)

Missing data was imputed using the MICE procedure and five imputed data sets were created. In each of the five imputed data sets the two step bootstrap model selection procedure as described above was applied. Information on model stability was provided by studying which combination of predictors occurred most frequently in 2500 data sets.

Internal validation

The apparent performance of a predictive model is typically better in the data set in which the model has been developed compared to its performance in another similar data set [32]. This phenomenon is called overoptimism. Using a n = 200 samples bootstrap procedure for internal validation [33] the performance of each developed model was tested in similar populations as in the derivation sample. This method was used to estimate the overoptimism of the derived models, and to adjust the measures of performance.

Model evaluation

Derived models were evaluated by comparing the model's composition (combination of predictors). Next several measures of predictive performance were considered. Discrimination refers to how well a model distinguishes between patients with and without persistent symptoms and is quantified by the c-index that, for binary outcomes, is identical to the area under the ROC curve (AUC) [34]. The c-index varies between 0.5 and 1, with 0.5 indicating no discrimination above chance and 1 indicating perfect discrimination. The agreement between predicted probabilities and observed probabilities is called calibration and was measured by computing the slope of the calibration plot (predicted probabilities against observed frequencies). Well-calibrated models have a slope of 1. As a measure of the explained variance Nagelkerke's R2 was computed.

Software

All analyses were performed using the R-statistics software (version 2.4.0). The R Design package was used for the CCA, MICE was used for the MI and additional routines were developed for applying the bootstrap.

Results

The baseline patient characteristics are listed in Table 1. After three months 517 patients (88%) returned the follow-up questionnaire. Subjects lost to follow-up were younger (mean difference of 7 years) and showed more often an acute onset (47% versus 37%). Due to non-response the percentage of missing data was largest for the outcome measures (shoulder disability 12.3% and shoulder pain intensity 12.9%). Other (baseline) variables had missing values within the range of 0 to 9.2%. The combination of missing values in CCA resulted in the exclusion of 24.7% (disability model) and 28.8% (pain intensity model) of participants. In the CCA 12 variables showed a univariable association with persistent disability and 16 with persistent pain, resulting in an EPV of 11.9 for pain intensity and 17 for shoulder disability. In the five imputation data sets the EPV varied between 19.1 and 19.6 for disability and between 13.5 and 13.8 for pain intensity. This means that the analyses were sufficiently powered (with a sufficient number of cases in the models) to reliably estimate the associations between predictors and outcome.

Model composition

For all presented models, the directions of the associations (i.e. regression coefficients) between the selected predictors and outcome were the same for both disability and pain (data not presented). Tables 2 and 3 show that for both measures of outcome, model composition was influenced by missing data (CCA vs. MI-5). When models were derived from imputed data, model composition diverged from the CCA model. For both measures of outcome predictors with lower predictive abilities in the CCA (i.e. rank order according to regression coefficient estimates) were not included in the MI-5 (e.g. concomitant lower extremity pain for shoulder disability and for pain intensity; sporting activities and higher physical workload). Predictors that were no part of the CCA model entered the MI-5 model for persistent shoulder disability (e.g. duration of complaints, somatisation, external locus of control and age) were included in the MI-5 model.
Table 2

Complete case and multiple imputed model compositions for the outcome measure persistent shoulder disability

missing valuesCCAMI-5
rankrank
persistent shoulder disability*72 (12.3%)
inability to perform daily activities8 (1.4%)14
shoulder complaints in the past year27 (4.6%)2
both shoulders afflicted0 (0%)33
concomitant lower back pain0 (0%)41
concomitant lower extremity pain0 (0%)5
more disability at baseline2 (0.3%)67
longer duration of complaints1 (0.2%)2
higher scores for somatisation3 (0.5%)5
higher scores for external locus of control33 (5.6%)6
older age0 (0%)8

CCA      - complete case analysis

MI -5     - multiple imputation using 5 imputation files

rank      - the order of appearance of predictors in the derived model arranged by their predictive ability (regression coefficient)

*           - outcome measure

Table 3

Complete case and multiple imputed model compositions for the outcome measure persistent shoulder pain intensity

missing valuesCCAMI-5
rankrank
persistent shoulder pain intensity*76 (12.9%)
sporting injury0 (0%)11
concomitant lower back pain0 (0%)23
longer duration of complaints1 (0.2%)32
both shoulders afflicted0 (0%)44
inability to perform daily activities8 (1.4%)55
concomitant upper extremity pain0 (0%)66
sporting activities0 (0%)7
higher physical workload0 (0%)8

CCA      - complete case analysis

MI -5     - multiple imputation using 5 imputation files

rank      - the order of appearance of predictors in the derived model arranged by their predictive ability (regression coefficient)

%          - inclusion frequency; the proportion of times that a variable with a univariable association with the outcome is retained in the automated backward selected models. When a variable was selected in each of the replications, the inclusion frequency was 100%

*           - outcome measure

Complete case and multiple imputed model compositions for the outcome measure persistent shoulder disability CCA      - complete case analysis MI -5     - multiple imputation using 5 imputation files rank      - the order of appearance of predictors in the derived model arranged by their predictive ability (regression coefficient) *           - outcome measure Complete case and multiple imputed model compositions for the outcome measure persistent shoulder pain intensity CCA      - complete case analysis MI -5     - multiple imputation using 5 imputation files rank      - the order of appearance of predictors in the derived model arranged by their predictive ability (regression coefficient) %          - inclusion frequency; the proportion of times that a variable with a univariable association with the outcome is retained in the automated backward selected models. When a variable was selected in each of the replications, the inclusion frequency was 100% *           - outcome measure Tables 4, 5, 6 and 7 show the results of assessing model stability by the bootstrap model selection procedure. CCA and MI-5 models were not identified as the most frequently occurring combination of predictors for both outcome measures (Tables 4, 5, 6). Only the persistent shoulder disability MI-5 method was identical to its bootstrapped enhanced version (Table 7). Model selection frequencies for the most frequently selected models were uniformly low (ranging from 24.0% to 3.6%). Indicating on a large variability in model composition within the bootstrap replicate data sets. When fewer potential predictors are retained after the first step of the bootstrap model selection procedure, this variability seemed to decrease and model selection frequency increased.
Table 4

Complete case bootstrap model selection for the outcome measure persistent disability

most frequently selected modelsrank
Predictors*12245BCCA
inability to perform daily activitiesXXXXX11
both shoulders afflictedX--XX23
shoulder complaints in the past yearXXXX-32
concomitant lower extremity painXXXXX45
more disability at baselineXXXXX56
concomitant lower back pain--XXX-4
older age-------
longer duration of complaints-------
acute onset-------
Count3323232216
%6.64.64.64.43.2

*            - only those predictors that appeared in ≥40% of the first bootstrap model selection step are presented

rank       - the order of appearance of predictors in the derived models arranged by their predictive ability (regression coefficient estimates)

B           - the complete case date based bootstrap selected model (i.e. the most frequently occurring combination of predictors in 500 replicate data sets of the second bootstrap model selection step)

CCA       - the complete case data based model derived without additional bootstrap was the fourth most occurring combination of predictors in the bootstrap model selection procedure

Count     - the number of times the model was selected in the 500 replicate data sets of the second bootstrap model selection step

Table 5

Complete case bootstrap model selection results for the outcome measure persistent pain intensity

most frequently selected modelsrank
Predictors12345BCCA
longer duration of complaintsXXXXX13
concomitant lower back painXXXXX22
both shoulders afflictedXX-XX34
concomitant upper extremity painXXX--46
shoulder complaints in the past yearX-XX-5-
sporting injury*-1
inability to perform daily activities*-5
sporting activities*-7
higher physical workload*-8
Count12096584737
%24.019.211.69.47.4

*             - predictors that appeared in ≥ 40% in the first step of the of the bootstrap model selection are not used in the second step in model selection

rank        - the order of appearance of predictors in the derived models arranged by their predictive ability (regression coefficient estimates)

B            - the complete case date based bootstrap selected model (i.e. the most frequently occurring combination of predictors in 500 replicate data sets of the second bootstrap model selection step)

CCA        - the complete case data based model derived without additional bootstrap did not occur in the bootstrap model selection since some of the included predictors occurred ≥ 40% in the first selection step

Count     - the number of times the model was selected in the 500 replicate data sets of the second bootstrap model selection step

Table 6

Imputed bootstrap model selection results for the outcome measure persistent pain intensity

most frequently selected modelsrank
Predictors*12345MI-5+BMI-5
sporting injuryXXXXX11
longer duration of complaintsXXXXX22
concomitant lower back painXXXXX33
both shoulders afflictedXXXXX44
inability to perform daily activitiesXXXXX55
higher level of educationXXX-X6-
shoulder complaints in the past yearXX--X7-
concomitant upper extremity painX-XXX86
higher physical workloadXXX-X9-
Count163158113111105
%6.56.34.54.44.2

*             - only those predictors that appeared in ≥40% of the first bootstrap model selection step are presented

rank        - the order of appearance of predictors in the derived models arranged by their predictive ability (regression coefficient estimates)

MI-5+B    - the multiple imputation based bootstrap selected model (i.e. the most frequently occurring combination of predictors in 2500 replicate data sets of the second bootstrap model selection step)

MI-5        - the multiple imputation based model using 5 imputed data sets was the fourth most occurring combination of predictors in the bootstrap model selection procedure.

Count      - the number of times the model was selected in the 2500 replicate data sets of the second bootstrap model selection step

Table 7

Imputed bootstrap model selection results for the outcome measure persistent disability

most frequently selected modelsrank
Predictors*12345MI-5+BMI-5
concomitant lower back painXXXXX11
longer duration of complaintsXXXX-22
both shoulders afflictedXXXXX33
inability to perform daily activitiesXXX-X44
higher scores for somatisationXXXXX55
higher scores for external locus of controlXXXXX66
more disability at baselineXXXXX77
older ageXX-XX88
shoulder complaints in the past year-XX-X--
concomitant lower extremity pain-------
Count9177565452
%3.63.12.22.22.1

*             - only those predictors that appeared in ≥40% of the first bootstrap model selection step are presented

rank        - the order of appearance of predictors in the derived models arranged by their predictive ability (regression coefficient estimates)

MI-5+B    - the multiple imputation based bootstrap selected model (i.e. the most frequently occurring combination of predictors in 2500 replicate data sets of the second bootstrap model selection step)

MI-5        - the multiple imputation based model using 5 imputed data sets was also the most frequently occurring combination of predictors in the 2500 bootstrap replicate data sets

Count      - the number of times the model was selected in the 2500 replicate data sets of the second bootstrap model selection step.

Complete case bootstrap model selection for the outcome measure persistent disability *            - only those predictors that appeared in ≥40% of the first bootstrap model selection step are presented rank       - the order of appearance of predictors in the derived models arranged by their predictive ability (regression coefficient estimates) B           - the complete case date based bootstrap selected model (i.e. the most frequently occurring combination of predictors in 500 replicate data sets of the second bootstrap model selection step) CCA       - the complete case data based model derived without additional bootstrap was the fourth most occurring combination of predictors in the bootstrap model selection procedure Count     - the number of times the model was selected in the 500 replicate data sets of the second bootstrap model selection step Complete case bootstrap model selection results for the outcome measure persistent pain intensity *             - predictors that appeared in ≥ 40% in the first step of the of the bootstrap model selection are not used in the second step in model selection rank        - the order of appearance of predictors in the derived models arranged by their predictive ability (regression coefficient estimates) B            - the complete case date based bootstrap selected model (i.e. the most frequently occurring combination of predictors in 500 replicate data sets of the second bootstrap model selection step) CCA        - the complete case data based model derived without additional bootstrap did not occur in the bootstrap model selection since some of the included predictors occurred ≥ 40% in the first selection step Count     - the number of times the model was selected in the 500 replicate data sets of the second bootstrap model selection step Imputed bootstrap model selection results for the outcome measure persistent pain intensity *             - only those predictors that appeared in ≥40% of the first bootstrap model selection step are presented rank        - the order of appearance of predictors in the derived models arranged by their predictive ability (regression coefficient estimates) MI-5+B    - the multiple imputation based bootstrap selected model (i.e. the most frequently occurring combination of predictors in 2500 replicate data sets of the second bootstrap model selection step) MI-5        - the multiple imputation based model using 5 imputed data sets was the fourth most occurring combination of predictors in the bootstrap model selection procedure. Count      - the number of times the model was selected in the 2500 replicate data sets of the second bootstrap model selection step Imputed bootstrap model selection results for the outcome measure persistent disability *             - only those predictors that appeared in ≥40% of the first bootstrap model selection step are presented rank        - the order of appearance of predictors in the derived models arranged by their predictive ability (regression coefficient estimates) MI-5+B    - the multiple imputation based bootstrap selected model (i.e. the most frequently occurring combination of predictors in 2500 replicate data sets of the second bootstrap model selection step) MI-5        - the multiple imputation based model using 5 imputed data sets was also the most frequently occurring combination of predictors in the 2500 bootstrap replicate data sets Count      - the number of times the model was selected in the 2500 replicate data sets of the second bootstrap model selection step.

Model performance

Table 8 presents the performance of the models derived with the four methods for both outcome measures. The slopes of the calibration plots ranged from 0.973 to 1.077, which indicates good calibration. Explained variance ranged from 8.8% to 12.0% for disability and from 13.5% to 18.8% for pain. The apparent c-indices varied between 0.645 and 0.667 for disability and between 0.684 and 0.717 for pain intensity. CCA models were more optimistic compared to the other models. Following adjustment for overoptimism the corrected c-indices were within the range of 0.639 - 0.646 for persistent shoulder disability and within the range of 0.667 - 0.688 for persistent shoulder pain.
Table 8

Model performance parameters.

Persistent disabilityPersistent shoulder pain intensity
CCAMI-5BMI-5 + BCCAMI-5BMI-5 + B
calibration slope0.9780.9781.0770.9780.9850.9730.9980.986
R2N0.1190.1200.0880.1200.1880.1620.1350.174
Ac 95% CI0.666 0.616,0.7150.667 0.624,0.7100.645 0.596,0.6940.667 0.624,0.7100.717 0.668,0.7660.702 0.660,0.7450.684 0.637,0.7320.710 0.668,0.752
Opt0.0270.0220.0230.0220.0300.0140.0180.022
Oc0.6390.6460.6220.6460.6860.6880.6670.688

Ac               - apparent c-index

B                - bootstrapping based on a complete case data set

CCA            - complete case analysis

MI-5+B        - multiple imputation combined with bootstrapping

MI-5            - multiple imputation using 5 imputation files

Oc              - optimism corrected c-index

Opt             - estimation of the overoptimism

R2N            - explained variance (Nagelkerke's R-squared)

95% CI        - 95% confidence interval

Model performance parameters. Ac               - apparent c-index B                - bootstrapping based on a complete case data set CCA            - complete case analysis MI-5+B        - multiple imputation combined with bootstrapping MI-5            - multiple imputation using 5 imputation files Oc              - optimism corrected c-index Opt             - estimation of the overoptimism R2N            - explained variance (Nagelkerke's R-squared) 95% CI        - 95% confidence interval

Discussion

Prognostic research aims at identifying the ingredients that contribute to a correct prognosis for a specific subgroup of patients. Though, finding a stable set of predictors that can consistently be used in a broad patient population proves to be difficult. Several methodological issues (missing data and model stability) which are not accounted for by the standard statistical methodology are expected to complicate this matter. We showed that accounting for missing data by MI and providing information on model stability by bootstrapping are instructive methods when deriving a prognostic model. In the standard statistical methodology the use of a backward or forward selection strategy has been criticized. It may result in overoptimistically estimated regression coefficients, omission of important predictors and random selection of less important predictors. Derived models may therefore be unstable. Research has focussed on how to derive stable models. One frequently used method is the bootstrapping approach suggested by Austin and Tu [35]. It considers the strength of evidence that identified variables are truly important predictors in re-sampled data. Although this approach is often claimed to reduce model instability [8,10,14,35,36], separating strong from weak predictors was shown to perform comparative to automated backward elimination in identifying the true regression model [37]. Furthermore, this approach has limited abilities when there is a high number of potential prognostic factors. For these situations a modified bootstrapping procedure was suggested [11]. Our study showed that the application of this two-step bootstrap model selection procedure provides valuable information on model stability. As frequently described, model size and model composition are also affected by missing data. Especially in standard statistical methodology where subjects with missing values on any of the recorded variables are omitted from analysis. When missing data does not depend on observed or unobserved measurements (Missing Completely At Random, MCAR), this leads to loss of costly gathered information, decreased statistical power, altered associations between predictors and therefore differences in model composition [12,13,38-41]. In this context our study findings formed no exception. Model composition varied as a result of whether cases with missing data were omitted from analyses (CCA) or whether the values of the missings were estimated using MI. Since missing values appeared to be related to other observed information, the MCAR condition did not hold and CCA was expected to be biased. Most of the missing data was observed in the outcome because participants did not consent to follow-up. As subjects lost to follow-up showed more often an acute onset (47% versus 37%), were younger (mean difference of 7 years) and the variable age is included in the MI model for the outcome measure persistent shoulder disability, it is plausible to assume that these missings are MAR. For that reason, accounting for missing data by MI using 5 imputed data sets was in our multivariate data setting the most optimal choice to reduce the uncertainty in model derivation caused by missing values. The use of even more data sets in the imputation routine is possible (up to 20), however 5 was shown to be an sufficient number in order to get stable results [30]. Yet the addition of a bootstrap model selection procedure showed that the MI-5 model might still be unstable. A possible source for this instability might be the suboptimal variable selection procedure applied in the MI-5 procedure. However, how to optimally perform variable selection in multiple imputed data is still a subject of discussion [42]. As illustrated by our study, the bootstrap model selection procedure may provide valuable additional information on model stability when deriving a prognostic model in multiple imputed data. To study the effects of accounting for missing data and incorporating model stability we used a large clinical data set in which we empirically evaluated different methods of deriving a prognostic model. By this, the uncertainties researchers commonly face when knowledge of the true predictors of outcome is lacking, were illustrated. Furthermore, the practical utility of the additional information provided by the bootstrap model selection procedure in prognostic modeling is demonstrated. Though results need to be interpreted with caution, as our approach limits us from identifying a superior methodology. Although performance parameters for each derived model are presented, these play no role in the decision on the superiority of a certain method. They only show that the performance of all derived models was comparable to that from existing clinical prediction rules on shoulder pain [6,15]. For deciding on the superiority of a certain method, a simulation study in which true predictors and noise variables are assigned would be needed. Such data is not presented by this study.

Conclusions

Our study showed that in this particular dataset of shoulder pain patients, model composition varied as a result of how missing data was handled. Furthermore, the bootstrap model selection routine gave additional information on model stability.

Competing interests

The authors declare that they have no competing interests.

Authors' contributions

Authors DvdW and TK were responsible for the conception, design and data collection of the study. Author MH designed the study's analytic strategy and helped to prepare the Material & Methods and the Discussion sections. Authors PC, GP, HvdH, HdV, MH, and DvdW critically appraised and revised draft versions of the manuscript. Author DV prepared the data for analyses, designed the analytic strategy, conducted all analyses and wrote the paper. All authors read and approved the manuscript.

Pre-publication history

The pre-publication history for this paper can be accessed here: http://www.biomedcentral.com/1471-2288/10/81/prepub
  32 in total

1.  Shoulder disability questionnaire design and responsiveness of a functional status measure.

Authors:  G J van der Heijden; P Leffers; L M Bouter
Journal:  J Clin Epidemiol       Date:  2000-01       Impact factor: 6.437

2.  Developing a prognostic model in the presence of missing data: an ovarian cancer case study.

Authors:  Taane G Clark; Douglas G Altman
Journal:  J Clin Epidemiol       Date:  2003-01       Impact factor: 6.437

3.  Stability of multivariable fractional polynomial models with selection of variables and transformations: a bootstrap investigation.

Authors:  P Royston; W Sauerbrei
Journal:  Stat Med       Date:  2003-02-28       Impact factor: 2.373

Review 4.  Systematic review of prognostic cohort studies on shoulder disorders.

Authors:  Ton Kuijpers; Daniëlle A W M van der Windt; Geert J M G van der Heijden; Lex M Bouter
Journal:  Pain       Date:  2004-06       Impact factor: 6.961

5.  Bootstrap model selection had similar performance for selecting authentic and noise variables compared to backward variable elimination: a simulation study.

Authors:  Peter C Austin
Journal:  J Clin Epidemiol       Date:  2008-06-09       Impact factor: 6.437

6.  A bootstrap resampling procedure for model building: application to the Cox regression model.

Authors:  W Sauerbrei; M Schumacher
Journal:  Stat Med       Date:  1992-12       Impact factor: 2.373

7.  A Fear-Avoidance Beliefs Questionnaire (FABQ) and the role of fear-avoidance beliefs in chronic low back pain and disability.

Authors:  Gordon Waddell; Mary Newton; Iain Henderson; Douglas Somerville; Chris J Main
Journal:  Pain       Date:  1993-02       Impact factor: 6.961

8.  A comparison of anlaytic methods for non-random missingness of outcome data.

Authors:  S L Crawford; S L Tennstedt; J B McKinlay
Journal:  J Clin Epidemiol       Date:  1995-02       Impact factor: 6.437

9.  Multiple regression for physiological data analysis: the problem of multicollinearity.

Authors:  B K Slinker; S A Glantz
Journal:  Am J Physiol       Date:  1985-07

10.  Evaluating the yield of medical tests.

Authors:  F E Harrell; R M Califf; D B Pryor; K L Lee; R A Rosati
Journal:  JAMA       Date:  1982-05-14       Impact factor: 56.272

View more
  13 in total

1.  Development and External Validation of a Mortality Prediction Model for Community-Dwelling Older Adults With Dementia.

Authors:  W James Deardorff; Deborah E Barnes; Sun Y Jeon; W John Boscardin; Kenneth M Langa; Kenneth E Covinsky; Susan L Mitchell; Elizabeth L Whitlock; Alexander K Smith; Sei J Lee
Journal:  JAMA Intern Med       Date:  2022-09-26       Impact factor: 44.409

2.  Smaller brained cliff swallows are more likely to die during harsh weather.

Authors:  Gigi S Wagnon; Charles R Brown
Journal:  Biol Lett       Date:  2020-07-15       Impact factor: 3.703

3.  Parasites favour intermediate nestling mass and brood size in cliff swallows.

Authors:  Charles R Brown; Mary Bomberger Brown
Journal:  J Evol Biol       Date:  2017-12-21       Impact factor: 2.411

4.  Prairie dogs increase fitness by killing interspecific competitors.

Authors:  John L Hoogland; Charles R Brown
Journal:  Proc Biol Sci       Date:  2016-03-30       Impact factor: 5.349

5.  Body mass index is not a clinically meaningful predictor of patient reported outcomes of primary hip replacement surgery: prospective cohort study.

Authors:  A Judge; R N Batra; G E Thomas; D Beard; M K Javaid; D W Murray; P A Dieppe; K E Dreinhoefer; K Peter-Guenther; R Field; C Cooper; N K Arden
Journal:  Osteoarthritis Cartilage       Date:  2014-01-11       Impact factor: 6.576

6.  Prediction of persistent shoulder pain in general practice: comparing clinical consensus from a Delphi procedure with a statistical scoring system.

Authors:  David Vergouw; Martijn W Heymans; Henrica C W de Vet; Daniëlle A W M van der Windt; Henriëtte E van der Horst
Journal:  BMC Fam Pract       Date:  2011-06-30       Impact factor: 2.497

7.  The estimation and use of predictions for the assessment of model performance using large samples with multiply imputed data.

Authors:  Angela M Wood; Patrick Royston; Ian R White
Journal:  Biom J       Date:  2015-01-29       Impact factor: 2.207

8.  Validation of prediction models based on lasso regression with multiply imputed data.

Authors:  Jammbe Z Musoro; Aeilko H Zwinderman; Milo A Puhan; Gerben ter Riet; Ronald B Geskus
Journal:  BMC Med Res Methodol       Date:  2014-10-16       Impact factor: 4.615

9.  The association of patient characteristics and surgical variables on symptoms of pain and function over 5 years following primary hip-replacement surgery: a prospective cohort study.

Authors:  Andy Judge; Nigel K Arden; Rajbir N Batra; Geraint Thomas; David Beard; M Kassim Javaid; Cyrus Cooper; David Murray
Journal:  BMJ Open       Date:  2013-03-01       Impact factor: 2.692

10.  Risk prediction in sexual health contexts: protocol.

Authors:  Titilola Falasinnu; Paul Gustafson; Mark Gilbert; Jean Shoveller
Journal:  JMIR Res Protoc       Date:  2013-12-03
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.