Literature DB >> 27097747

Outcome modelling strategies in epidemiology: traditional methods and basic alternatives.

Sander Greenland1, Rhian Daniel2, Neil Pearce3.   

Abstract

Controlling for too many potential confounders can lead to or aggravate problems of data sparsity or multicollinearity, particularly when the number of covariates is large in relation to the study size. As a result, methods to reduce the number of modelled covariates are often deployed. We review several traditional modelling strategies, including stepwise regression and the 'change-in-estimate' (CIE) approach to deciding which potential confounders to include in an outcome-regression model for estimating effects of a targeted exposure. We discuss their shortcomings, and then provide some basic alternatives and refinements that do not require special macros or programming. Throughout, we assume the main goal is to derive the most accurate effect estimates obtainable from the data and commercial software. Allowing that most users must stay within standard software packages, this goal can be roughly approximated using basic methods to assess, and thereby minimize, mean squared error (MSE).
© The Author 2016. Published by Oxford University Press on behalf of the International Epidemiological Association.

Entities:  

Mesh:

Year:  2016        PMID: 27097747      PMCID: PMC4864881          DOI: 10.1093/ije/dyw040

Source DB:  PubMed          Journal:  Int J Epidemiol        ISSN: 0300-5771            Impact factor:   7.196


The main goal of a statistical analysis of effects should be the production of the most accurate (valid and precise) effect estimates obtainable from the data and available software. This goal is quite different from that of variable selection, which is to obtain a model that predicts observed outcomes well with the minimal number of variables; this prediction goal is only indirectly related to the goal of change-in-estimate approaches, which is to obtain a model that controls most or all confounding with a minimal number of variables. We illustrate some basic alternative modelling strategies that focus more closely on accurate effect estimation as measured by mean squared error (MSE) and which can be implemented by practitioners with limited programming and consulting resources.

Introduction

We have recently reviewed traditional approaches to confounder selection for outcome (risk) and treatment (propensity) models, including significance-testing and ‘change-in-estimate’ (CIE) approaches. We argued that the main goal of a statistical analysis of effects should be the production of the most accurate (valid and precise) effect estimates obtainable from the data and available software. Allowing that most users must stay within standard software packages, this goal can be roughly approximated using basic methods to minimize estimated mean squared error (MSE). We here provide an illustrated overview of this approach.

Scope, aims and assumptions

As with our initial review, our coverage is not intended for highly skilled practitioners; rather, we target teachers, students and working epidemiologists who would like to do better with data analysis, but who lack resources such as R programming skills or a bona fide modelling expert committed to their project. Throughout, we assume that we are applying a conventional risk or rate regression model (e.g. logistic, Cox or Poisson regression) to estimate the effects of an exposure variable X on the distribution of a disease variable Y while controlling for other variables, and that the outcome is uncommon enough so that distinctions among risk, rate and odds ratios can be ignored. The other variables include forced variables, such as age and sex, which we may always want to control, and may also include unforced variables about which we are unsure whether to control. We also assume that data checking, description and summarization have been done carefully. Finally, we assume that all quantitative variables have been: re-centreed to ensure that zero is a meaningful reference value present in the data; and rescaled so that their units are meaningful differences within the range of the data; and that univariate distributions and background (contextual) information have been used to select categories or an appropriately flexible form (e.g. splines) for detailed modelling. Elsewhere we have discussed the issues involved in simply adjusting for all measured potential confounders. This approach can be valid when the number of covariates is not too large in relation to the study size and the included covariates are not highly predictive of exposure. Nonetheless, controlling too many variables can lead to or aggravate problems arising from data sparsity or from high multiple correlation of exposure with the controlled confounders (which we term multicollinearity), in which case one may seek to reduce the number of modelled covariates. There are of course variables for which control may be inappropriate based on preliminary causal considerations. These include intermediates (variables on the causal pathway between exposure and diseases) and their descendants and any other variable influenced by the exposure or outcome. These also include variables that are not part of minimal sufficient adjustment sets, whose control may increase bias. We assume that these variables have been identified and eliminated e.g. using causal diagrams ,, to display contextual theory, leaving us with a set of potential adjustment covariates (often called ‘potential confounders’), including those variables that we are reasonably confident would reduce bias if controlled and our study size were unlimited. We focus only on basic selection from these variables, leaving aside many difficult issues about model specification and diagnostics, , time-varying exposures and confounders, interactions and mediation.

Multicollinearity and mean squared error: modified CIE approaches

One issue that is not explicitly considered or discussed in most epidemiological strategies is that of multicollinearity of covariates with exposure, i.e. when exposure is nearly a linear combination of other variables in the model. This problem becomes most obvious in propensity-score analyses when the exposure is so well predicted that there is little overlap in the exposed and unexposed scores. With multicollinearity, exposure effect estimates become unstable, as reflected by large standard errors. To combine bias and variance considerations when dealing with genuine confounders, consider estimation of an exposure effect measure represented by a single coefficient β , such as a rate difference or log risk ratio. The bias B in an estimator of β is the difference between the expected value (mean) μ of the estimator and the ‘true’ population value β , so B =  μ – β . The standard error (SE) of the estimator is just its standard deviation around that mean μ ; SE 2 is thus the estimator’s variance. The mean squared error (MSE) of the estimator of β combines these properties via the equation MSE = B 2 + SE 2 . Reducing multicollinearity by dropping variables can decrease the variance (SE 2 ) component of the MSE, but may also increase the bias B in the estimator of β if the dropped variables are indeed necessary to adjust for, given the retained variables. Thus we seek ways of reducing the SE of the estimator (e.g. by removing a source of multicollinearity) without seriously increasing its bias B, so that the MSE is reduced. ,, Several formal methods seek to minimize MSE in effect estimation with uncertain confounders, but require special programming. ,, We will describe a more crude approach that extends ordinary CIE approaches to consider estimated MSE minimization using ordinary software outputs. Suppose we selectively delete confounders from a full model and see what happens to the exposure coefficient estimate and its standard error. Assuming the full-model estimate is unbiased, we can then estimate the bias B reduced from the deletion by the difference between the reduced-model estimate reduced and full-model estimate full . This step leads to the following equations for estimating the change in MSE (ΔMSE) from reducing the model by deleting the confounder: where (ΔB) estimates the squared-bias increase from the deletion and Δ(SE 2 ) estimates the variance decrease from the deletion. A positive difference, i.e. (ΔB) > Δ(SE 2 ), indicates that the deletion increased the MSE; a negative difference indicates that the deletion reduced the MSE. We say ‘indicates’ because, of course, we have only rough estimates of B, SE and MSE, and full , which will be approximately unbiased only when the model, the set of measured confounders and the sample size are all sufficient for approximate validity. This approach is illustrated in Box 1, with an example involving two correlated variables, sodium and potassium intake. Box 1 We consider an example from a study of sodium intake in infancy (age 4 months) and blood pressure at 7 years. The analysis involved adjusting for a relatively large number of potential confounders (see Table 1 ). A potentially important confounder was potassium intake at the same age, which was strongly correlated with sodium intake (r = 0.81). This was reflected in an increase in the standard error for the sodium coefficient when potassium was also included in the model. The authors therefore note that ‘due to high sodium-potassium correlations, effect of sodium independent of potassium could not be estimated with reasonable precision’, and they therefore did not control for potassium in the analyses.
Table 1.

Associations of sodium and potassium intake at age 4 months with blood pressure (BP) at age 7 years

Model Exposure variables *Coefficient estimateSE for coefficientCoefficient bias estimateIndicates biasIndicates large collinear Root MSE estimate *
1Sodium0.5180.290Referent0.290
Potassium0.0990.095Referent0.095
2aSodium0.7080.2250.190YesYes0.294
2bPotassium0.2060.0740.107YesYes0.130

*All analyses are adjusted for energy intake at 4 or 8 months, age at BP measurement, sex, socioeconomic position (maternal and paternal education), family social class, maternal age at childbirth, parity, birthweight, gestational age, breastfeeding, smoking during pregnancy, sodium intake at 7 years.

We did RMSE analyses ( Table 1 ), which showed that although there was an increase in the SE of the sodium coefficient when potassium is included in the model (compare model 1 with model 2a), the reduction in SE from deleting potassium from the model is offset by the increase in bias (sodium RMSE = 0.294 with potassium excluded vs 0.290 with potassium included). Thus, controlling for potassium appears to be no worse in accuracy, in addition to having smaller approximate bias. Next, consider potassium as the main exposure: we obtain a lower RMSE (0.095) for the potassium coefficient when including sodium compared with excluding sodium (0.130); thus controlling for sodium appears to be preferable. Associations of sodium and potassium intake at age 4 months with blood pressure (BP) at age 7 years *All analyses are adjusted for energy intake at 4 or 8 months, age at BP measurement, sex, socioeconomic position (maternal and paternal education), family social class, maternal age at childbirth, parity, birthweight, gestational age, breastfeeding, smoking during pregnancy, sodium intake at 7 years. As with CIE, the exposure-coefficient change resulting from covariate deletion can be assessed by examining the estimated change directly, and also with a collapsibility test, i.e. a test of the hypothesis that the deletion does not change the exposure coefficients. One caution to these approaches is that an accurate assessment of confounding may require examining changes from moving groups of variables. Regardless of the number of covariates being deleted, however, if there is one exposure term X, then a one degree of freedom chi-squared statistic for this hypothesis is χ c  = (ΔB) 2 /Δ(SE 2 ). Deleting a variable when ΔMSE > 0 is equivalent to deleting the variable when χ c < 1, which corresponds to P > 0.32 for collapsibility. Appendix 1 (available as Supplementary data at IJE online) gives further details, describes a generalization of this test to exposures represented by multiple terms and suggests avenues for improvement. To illustrate the general algorithms, denote by W 1 ,…,W J those variables (such as age and sex) that we want forced into all our models along with exposure X because they are expected to be important confounders or modifiers of the exposure effect measure, or because they are known strong risk factors that everyone wants to see in adjustment; this list could include age splines, sex and ethnicity indicators etc. Our chief concern will be with the remaining variables U 1 ,…,U H , whose importance for adjustment is highly uncertain. Some hypothetical modelling results are shown in Table 2 . We suppose result 1 is from a full model for the disease rate with exposure, the forced variables and all potential confounders. Results 2a–d then illustrate the four mutually exclusive possible outcomes of comparing a full (maximal) model including the potential confounders (forced and unforced variables) with a minimal model including only the main exposure and the forced variables. Result 2a suggests little or no confounding or multicollinearity problems, since there is little difference between the basic and full models; we might therefore prefer the simplicity of reporting estimates from the minimal model. In contrast, result 2b suggests there is confounding by the unforced variables, as seen by contrasting the exposure rate ratios from model 1 and model 2b, indicating that it is necessary to control at least some of the unforced variables.
Table 2.

Hypothetical results from rate regressions in which a covariate is or is not a confounder or a source of multicollinearity

ModelModel variablesExposure coefficient estimateRate ratio estimateSE for coeff.95% CL Coefficient bias estimate *Indicates bias?Indicates strongly collinear? Root MSE estimate * Collapsibility χ 2 and P -value 33
1 X,W 1 …W J , U 1 …U H0.6932.000.241.25,3.20Referent0.24
Some mutually exclusive alternative possibilities under model 2 (minimal model in which all unforced variables U 1 …U H are dropped)
2a X,W 1 …W J0.6932.000.241.25, 3.200NoNo0.24 0, P = 1
2b X,W 1 …W J1.0993.000.202.03, 4.440.405YesNo0.45 9.34, P = 0.002
2c X,W 1 …W J0.6932.000.141.52, 2.630NoYes0.14 0, P = 1
2d X,W 1 …W J1.0993.000.142.28, 3.950.405YesYes0.43 4.03, P = 0.04

*Taking model 1 as the referent (‘gold standard’).

Hypothetical results from rate regressions in which a covariate is or is not a confounder or a source of multicollinearity *Taking model 1 as the referent (‘gold standard’). Results 2c and 2d involve large multicollinearity, as indicated by the difference (0.14 compared with 0.24) in the standard error for the main exposure coefficient. The more favourable situation is when the factors causing multicollinearity are very weak confounders, so they can be deleted from the model without increasing the MSE of the exposure-effect estimate. This situation is indicated when deleting these factors leaves the exposure-effect estimate virtually unchanged, but greatly reduces its standard error (as in result 2c), suggesting that the minimal model provides more accurate estimates of the exposure effect (i.e. it has a smaller MSE). Again, we caution that this smaller standard error does not account for the preliminary testing and is thus too small by an unknown amount. It is more difficult to proceed when multicollinearity arises from a strong confounder (result 2d), since the increase in precision due to deleting such a confounder may be more than offset by an increase in confounding. We thus must consider the net impact of reducing the SE of the exposure-effect estimate while increasing its bias, and we do so by directly comparing square roots of estimated MSE (RMSE); we use the square roots to put the results back on the scale of the effects and biases. In result 2d, the estimated RMSE from the minimal model is substantially larger (0.43) than from the full model (0.24), because the minimal model involves a large increase in confounding and a relatively smaller decrease in multicollinearity. The task is then to identify a compromise model (including some but not all the variables in question) in which multicollinearity is reduced, but there is negligible increase in confounding. This could occur, for example, if the variables most responsible for confounding were distinct from the variables most responsible for multicollinearity. Candidate variables can be assessed by dropping each variable in turn from the full model. Of course, this process may fail to identify any acceptable model reduction, in which case the options are to stay with the full model or else turn to more sophisticated methods such as penalized estimation or hierarchical (multilevel or mixed) models to improve accuracy. , Table 1 gives effect estimates without and with adjustment for the U h , which provides a basis for discussing the plausibility of residual confounding. For example, if adjustment using imperfectly measured U h removes more than one-half of the excess rate associated with a particular main exposure, then it is reasonable to speculate that adjustment with better U h information would have removed most of the excess rate. Thus it can be worthwhile to present estimates from different degrees of adjustment. Based on the above considerations, Box 2 outlines one backward-deletion strategy for screening out potential confounders. This strategy is intended as a set of options, rather than a prescription; it would be applicable in settings in which a full model can be fit without problems, there is not an inordinate number of potential confounders to consider and there is no clear and strong heterogeneity. One implementation is as follows: B1) Fit the full model, with no exposure-covariate products. This model provides an average regression across the included covariates, even if heterogeneity is present. B2) Enter the following reduction loop, starting with the full model as the ‘current model’: For each candidate variable that remains in the current model, re-run the model without its terms (the U h that represent it) and compute the resulting ΔMSE relative to the current model from dropping those terms; again, If any candidate in the model has ΔMSE < 0 (indicating its deletion reduces MSE), drop the one with the smallest (most negative) ΔMSE and go to step (a) if there is any candidate left in the model. Otherwise (if there is no candidate U h left in the model, or none left have ΔMSE < 0), stop and use the current model. Box 2 Variable selection based on backward deletion using estimated MSE reduction Baseline specification 1.1 Select the variables that are appropriate to include, using a causal directed acyclic graph (DAG) to exhibit theorized causal relations among variables identified a priori as potentially important for estimating the effects of interest. 1.2 Divide the variables into three classes: (i) the main exposure X; (ii) forced-in variables (e.g. age, sex) which are always included in the model (W 1 …W J ); and (iii) the non-forced variables which will be candidates for deletion (U 1 …U H ). 1.3 Run a ‘full’ model including all main exposure terms, forced-in variables and non-forced variables from 1.3, with no exposure-covariate products. [If full model does not converge or the results indicate sparse-data bias, change to a forward-selection strategy, or use hierarchical (multilevel or mixed) or penalized modelling methods.] Variable selection Enter the following reduction loop, starting with the full model as the ‘current model’: 2.1 For each candidate variable that remains in the current model, re-run the model without its terms (the U h that represent it)and compute the resulting ΔMSE relative to the current model from dropping those terms: 2.2 If any candidate has ΔMSE < 0, drop the one with the smallest (most negative) ΔMSE and go to step 4.2 if there are any candidates left in the model. Otherwise (if there is no candidate U h left in the model, or none left have ΔMSE < 0), stop and use the current model. Assessment of heterogeneity (effect-measure modification) 3.1 Assess heterogeneity in a series of supplementary analyses, focusing on covariates of a priori interest We can also derive a parallel forward-selection strategy starting with the basic model when there are more potential confounders to consider than can reasonably fit at once (e.g. when using too many of them results in sparse-data bias, thus spuriously inflating (ΔB) ): F1) Fit the basic model, with no exposure-covariate products. F2) Enter the following expansion loop, starting with the basic model as the ‘current model’: For each candidate variable that is not in the current model, re-run the model expanded with its terms U h and compute the ΔMSE from adding those terms. If any candidate U h not in the model has ΔMSE > 0 (indicating its addition reduces MSE), enter the one with the largest ΔMSE and go to step (a) if any candidate remains left out. Otherwise (if there are no more unselected candidates, or if none left out have ΔMSE > 0), stop and use the current model. Both the above approaches can be viewed as a modification of conventional testing strategies in one major way: the test of the confounder coefficient is replaced by a test of collapsibility of the exposure coefficient over the confounder. This test is easily constructed from ordinary outputs (see Appendix 1, available as Supplementary data at IJE online) and is appropriately sensitive to the confounder relation to exposure as well as to its relation to disease. It can also be viewed as a modification of CIE strategy that allows for random error in the observed change and for the possible variance reduction from deletion. In Box 3, these approaches are applied to a study of atopy in Poland, and their results are compared with other common approaches. Box 3 We consider an example from a study of the prevalence of atopy in a small town and neighbouring villages in Poland in 2003. In the current analysis, we estimate the association between ‘no current unpasteurized milk consumption’ and current atopy status. It was plausible that lack of unpasteurized milk consumption could increase the risk of atopy. Because drinking unpasteurized milk happens mostly in rural settings, however, there are a number of other exposures which may be related to both unpasteurized milk consumption and the prevalence of atopy. Main exposure: never drinking unpasteurized milk (1: never vs 0: regularly/sometimes). Forced variables: age-group (seven categories), sex. Potential confounders: Live in town (yes/no) or village Live on a farm (yes/no) Contact (regular/occasional) with cows, pigs, poultry, sheep or goats, horses Work (regular/occasional) milking cows, cleaning barns, collecting eggs Firstborn (yes/no) Number of siblings (1, 2, 3+) Current smoker (yes/no) Lived in town (yes/no) or village as a child Lived on a farm (yes/no) as a child Parents were farmers (yes/no) Family kept cows, pigs, poultry, sheep or goats, horses. Basic model Model 1 in Table 3 shows the results of the basic analysis for milk, adjusted for the forced variables (age-group and sex).
Table 3.

Model-adjusted associations of current unpasteurized milk consumption with current atopy status

Model Model variables *Exposure coefficient estimateSE for coefficientOR95% CL for OREstimated bias and RMSE Bootstrap SE for coefficient Bootstrap 95% CL for OR
1 (basic)Milk0.8990.2252.461.58, 3.82

0.516

0.567

0.2361.59, 3.97
2a (ML full)

Milk

All other variables #

0.4060.2571.500.91, 2.48

0.023

0.262

0.2610.89, 2.46
2b (Firth) 42,43

Milk

All other variables #

0.3830.2521.470.91, 2.40

0.000

0.251

0.2510.89, 2.37
3a (forwards stepwise, P < 0.20)

Milk

Town

Firstborn

Current smoker

Town as a child

Parents farmers

Parents kept poultry

Parents kept horses

0.3900.2441.480.91, 2.38

0.007,

0.261

0.2610.87, 2.43
3b (forwards stepwise, P < 0.05)

Milk

Town

Current smoker

Town as a child

Parents kept poultry

0.3830.2431.470.91, 2.36

<0.001

0.261

0.2610.88, 2.44
3c (backward stepwise, P < 0.20)

Milk

Town

Firstborn

Current smoker

Parents farmers

Parents kept poultry

Parents kept horses

0.3980.2441.490.92, 2.40

0.015

0.261

0.2610.88, 2.47
3d (backward stepwise, P < 0.05)

Milk

Town

Current smoker

Parents farmers

Parents kept poultry

Parents kept horses

0.4140.2441.510.94, 2.44

0.031

0.265

0.2630.93, 2.61
4a (forwards AIC)

Milk

Town

Horses

Firstborn

Current smoker

Parents kept poultry

0.3810.2431.460.91, 2.36

−0.002

0.260

0.2600.86, 2.39
4b (backward AIC)

Milk

Town

Horses

Firstborn

Current smoker

Parents kept poultry

Parents kept horses

Parents farmers

0.3980.2441.490.92, 2.40

0.015

0.262

0.2620.88, 2.48
5a (forwards BIC)

Milk

Town

Current smoker

Parents kept poultry

0.3930.2431.480.92, 2.39

0.010

0.264

0.2640.88, 2.45
5b (backward BIC)

Milk

Town

Current smoker

Parents kept poultry

0.3930.2431.480.92, 2.39

0.010

0.264

0.2640.87, 2.45
6a (forwards CIE)

Milk

Town

0.4000.2421.490.93, 2.39

0.017

0.255

0.2540.93, 2.56
6b (backward CIE)

Milk

Town

0.4000.2421.490.93, 2.39

0.017

0.255

0.2540.92, 2.52
7a (forwards RMSE, larger model as referent)

Milk

Town

Poultry

Collecting eggs

Number of siblings

Parents kept cows

Parents kept poultry

0.3630.2451.440.89, 2.32

−0.020

0.258

0.2570.86, 2.35
7b (backward RMSE, larger model as referent)

Milk

Town

Poultry

Collecting eggs

Firstborn

0.3500.2431.420.88, 2.29

0.017

0.257

0.2560.84, 2.28
8a (forwards RMSE, full model as referent)

Milk

Town

0.4000.2421.490.93, 2.39

−0.033

0.263

0.2610.88, 2.45
8b (backward RMSE, full model as referent)

Milk

Town

Parents kept cows

Parents kept poultry

0.4070.2421.500.94, 2.42

0.024

0.264

0.2630.89, 2.51
9a penalization by log-F(1,1) priors §45

Milk

All other variables #

0.3960.2531.490.90, 2.44

0.013

0.253

0.2530.90, 2.42
9b penalization by log-F(2,2) priors 45

Milk

All other variables #

0.3890.2501.470.90, 2.41

0.006

0.246

0.2460.90, 2.36

*All analyses are adjusted for age group and sex.

† Based on 4000 bootstrap samples.

‡ Bias-corrected and accelerated (BCa) with 4000 resamples.

# Town, farm, cows, pigs, poultry, sheep/goats, horses, milking cows, cleaning barns, collecting eggs, firstborn, number of siblings, current smoker, lived in town or village as a child, parents were farmers, family kept cows, family kept pigs, family kept poultry, family kept sheep or goats, family kept horses.

§ Equivalent to F(1,1) prior for odds ratio; 95% prior limits are 1/648, 648.

¶ Equivalent to F(2,2) prior for odds ratio; 95% prior limits are 1/39, 39.

Full model Model 2a in Table 3 shows the results of the full maximum likelihood (ML) model, adjusting for all potential confounders; there is a substantial change in the odds ratio for milk (from 2.46 to 1.50), but there is also an increase in the SE for the coefficient estimate (from 0.225 to 0.257). Model 2b is the full model fit using the Firth adjustment for coefficient-estimate bias. , This is used as the ‘standard’ to estimate the bias of the other models, and is combined with the bootstrap SEs to estimate the RMSE. Overall, the milk coefficients from the full models have a much lower RMSE (0.262, 0.251) than in the basic model (0.567) because the increase in SE from including all potential confounders is small in comparison with the change in the coefficient estimate. Traditional stepwise regression Model 3a in Table 1 shows the results of a forwards stepwise logistic regression (using P < 0.20 as the criterion for inclusion) with milk, age group and sex as forced variables; Town, Firstborn, Current smoker, Town as a child, Parents farmers, Parents kept poultry and Parents kept horses were also selected. Model 3b is again a forwards stepwise logistic regression but uses P < 0.05 as the criterion for inclusion. Model 3c and d are the backwards stepwise procedures with P < 0.20 and P < 0.05, respectively. AIC Model 4a in Table 1 shows the results of using the Akaike Information Criterion (AIC) where variables were forward selected to achieve the largest increase in AIC at each step. Model 4b is from using AIC for backwards deletion. BIC Model 5a and b was selected in parallel to 4a and b but using the Bayesian Information Criterion. Relative change-in-estimate approach Only town residence (in addition to the forced variables of age group and sex) produced a substantial change in the estimate for milk; once this was in the model, no other variable changed the milk odds ratio estimate by more than 10%, leading to model 6a. Model 6b is from the analogous backwards procedure and resulted in the same model. RMSE Model 7a in Table 1 shows the results of using RMSE reduction for forward selection in two different ways. Model 7a1 used (at each step) the larger of the two models being compared as the reference for estimating RMSE reduction, and is thus analogous to the other procedures, whereas model 7a2 used the full model as the reference for each step. Model 7b is the backwards version of the same procedure. Model 7b1 used (at each step) the larger of the two models being compared as the reference (for estimating the RMSE), whereas model 7b2 used the full model as the reference for each step. Penali z ation Following previous recommendations, , we included two analyses with weakly informative shrinkage priors for each coefficient. The first analysis used a log-F(1,1) (Haldane) prior distribution for each coefficient, which is equivalent to using an F(1,1) prior distribution for the odds ratio (antilog) from each coefficient, and assigns 95% probability to the odds ratio falling between 1/648 and 648. The second analysis used a log-F(2,2) (standard logistic) distribution for each coefficient, which is equivalent to using an F(2,2) prior distribution for the odds ratio from each coefficient, and assigns 95% probability to the odds ratio falling between 1/39 and 39. The priors were imposed by adding two pseudo-observations for each coefficient to the actual data file, with weights of ½ for the F(1,1) prior and weights of 1 for the F(2,2) prior, then fitting the full model to the augmented data set by maximum likelihood, with the constant term replaced by an indicator for ‘actual-data record’ and weights of 1 for all actual-data records. ,, Discussion In this example, all of the modelling approaches yielded reasonably similar findings—the full model (Firth bias-adjusted) yielded an OR of 1.47, and all of the other approaches produced ORs in the range of 1.42 to 1.51. The RMSEs were also similar, smaller than that of the full model and substantially smaller than that for the basic model. The fact that there exist models with lower estimated RMSE than the models selected by the RMSE procedures 7ab (using the larger of the two models as the reference) illustrates how a procedure that selects or rejects variables one at a time (forwards or backwards) does not always find the model with the overall optimal value of the criterion being used. In this example, Town is the only variable whose inclusion/exclusion in the model has much impact on the exposure effect estimate. Town is also highly predictive of the outcome. Thus, all methods select it, and whatever else they happen to select makes very little difference for any of the measures considered. For the same reasons, the bootstrap 95% CIs (which take variable selection into account) were in general only slightly larger than the ‘standard’ 95% CIs. We therefore see little apparent advantage of one method over another in this example. Nonetheless, in a setting with strong confounding by intercorrelated groups of multiple confounders, we might find more stark differences among the results from different methods. Model-adjusted associations of current unpasteurized milk consumption with current atopy status 0.516 0.567 Milk All other variables 0.023 0.262 Milk All other variables 0.000 0.251 Milk Town Firstborn Current smoker Town as a child Parents farmers Parents kept poultry Parents kept horses 0.007, 0.261 Milk Town Current smoker Town as a child Parents kept poultry <0.001 0.261 Milk Town Firstborn Current smoker Parents farmers Parents kept poultry Parents kept horses 0.015 0.261 Milk Town Current smoker Parents farmers Parents kept poultry Parents kept horses 0.031 0.265 Milk Town Horses Firstborn Current smoker Parents kept poultry −0.002 0.260 Milk Town Horses Firstborn Current smoker Parents kept poultry Parents kept horses Parents farmers 0.015 0.262 Milk Town Current smoker Parents kept poultry 0.010 0.264 Milk Town Current smoker Parents kept poultry 0.010 0.264 Milk Town 0.017 0.255 Milk Town 0.017 0.255 Milk Town Poultry Collecting eggs Number of siblings Parents kept cows Parents kept poultry −0.020 0.258 Milk Town Poultry Collecting eggs Firstborn 0.017 0.257 Milk Town −0.033 0.263 Milk Town Parents kept cows Parents kept poultry 0.024 0.264 Milk All other variables 0.013 0.253 Milk All other variables 0.006 0.246 *All analyses are adjusted for age group and sex. † Based on 4000 bootstrap samples. ‡ Bias-corrected and accelerated (BCa) with 4000 resamples. # Town, farm, cows, pigs, poultry, sheep/goats, horses, milking cows, cleaning barns, collecting eggs, firstborn, number of siblings, current smoker, lived in town or village as a child, parents were farmers, family kept cows, family kept pigs, family kept poultry, family kept sheep or goats, family kept horses. § Equivalent to F(1,1) prior for odds ratio; 95% prior limits are 1/648, 648. ¶ Equivalent to F(2,2) prior for odds ratio; 95% prior limits are 1/39, 39.

Some limitations

As with most variable-selection procedures including stepwise and CIE, confidence intervals obtained by combining the final point estimate and SE from the above strategy are not theoretically valid. Simulation studies , so far suggest that this invalidity is negligible in typical settings, due to the high significance level and therefore liberal inclusion implicit in using ΔMSE = 0 as the decision point. Nonetheless, the strategy could be improved by using bootstrapping or cross-validation to estimate ΔMSE and set confidence intervals. A further problem with using CIE strategies for logistic regression is that it is possible the change in estimate is largely due to more sparse-data bias (i.e. too few subjects at crucial combinations of variables) in the full-model estimate full rather than increased confounding in the reduced-model estimate reduced . For a binary exposure X and disease Y, this problem becomes noticeable when there are much fewer than about 4 subjects per confounder coefficient at each exposure-disease combination; for example, with 7 confounder terms we would want at least 4(7) = 28 subjects in each cell of the two-way XY table for some assurance that sparse-data bias in full is small. One way to avoid this problem is to switch to penalized estimation; it is also possible to apply the above reduction algorithms after minimal penalization to reduce sparse-data bias. Another problem however is that logistic coefficients are in general not collapsible, in that there will be differences between the actual (underlying) coefficients with and without a given covariate if the covariate predicts the outcome, even if that covariate is not a confounder by virtue of being independent of exposure. This difference will be negligible unless the outcome is common, in which case it will be advisable to switch to estimation of collapsible effect measures (such as risk ratios and differences), e.g. by regression standardization.

Discussion

Like more sophisticated but computationally intensive methods, the strategies we describe differ from stepwise regression and other purely predictive approaches, in that their goal is to improve accuracy of exposure effect estimates rather than to simply predict outcomes. At the same time, recognizing that the gap between state-of-the-art methods and what is done in most publications has only grown over time, they are intended to fall within the scope of the limits on software and effort that constrain typical researchers. Thus, parsimony is replaced by the goal of minimizing error in effect estimation. A related point is that, as with parsimony, pursuit of goodness-of-fit may lead to inappropriate decisions about confounder control; in particular, some variables may not be included in the model because they do not significantly improve the fit, even though they are important confounders. ‘Global’ tests of fit are especially inadequate for confounder selection since there can be many ‘good-fitting’ models that correspond to very different confounder effects and exposure effect estimates. Parsimony and goodness-of-fit are helpful only to the extent they reduce variance and bias of the targeted effect estimate. The general inappropriateness of parsimony as a goal in causal analysis is supported by simulation studies in which full-model analysis has often outperformed conventional selection strategies. ,, This result raises the question: if we can control for all potential confounders, then why wouldn’t we? If indeed we have numbers so large that there is no problem from controlling too many variables, we would generally expect covariate elimination to provide little benefit for the accuracy of effect estimates. But the harsh reality is that even databases of studies with hundreds of thousands of patients often face severe limits in crucial categories, such as the number of exposed cases. Coupled with the availability of what may be hundreds or even thousands of variables, some kind of algorithmic approach to potential confounders becomes essential. , The strategies we describe are designed for common borderline situations in which control of all the variables may be possible, but some accuracy improvement may be expected from eliminating some or all variables whose inclusion is of uncertain benefit. A number of criticisms can be made of the MSE-based strategy in Box 2. First, it can be argued that any data-based model reduction will produce biased estimates because it depends on the assumption that it is not necessary to control the omitted variables (conditional on control of the included variables). We regard this criticism as somewhat misguided insofar as every epidemiological estimate suffers from some degree of bias from uncontrolled confounders, differential subject selection and measurement error (in both exposures and confounders); the key question is then whether the bias from omitting a variable is of contextual importance. Second, as we have emphasized, simple selection methods (such as stepwise, CIE and apparent MSE change) do not take account of random variability introduced by data-based model selection. Thus, without cross-validation or some other adjustment, the standard error of the resulting effect estimate is not correctly estimated by taking the standard error computed from the final model. With methods that focus on the effect estimate, however, the eliminated variables are generally those that have only weak relations to exposure or disease, the resulting problem is limited. Where such problems are of concern, they can be mitigated by the use of shrinkage, penalization and related hierarchical methods, ,,,,,, model averaging, , cross-validation or bootstrapping. Third, the MSE approaches we describe may encounter technical difficulties in precisely the situation of most concern here, namely when there is multicollinearity. As we mentioned, sparse-data bias is a chief concern along with related artefacts due to sample-size limitations, which again suggests using in the MSE algorithms the bias-reduced estimates available in commercial software. , The strategies we have presented in this paper are in no sense optimal; rather they are rough but transparent heuristics which attempt to mitigate some of the difficulties of common approaches without introducing too much new machinery or subtle statistical concepts. Regardless of the strategy adopted, however, it is important that authors document how they chose their models, so that readers can interpret their results in light of the strengths and weaknesses attendant on the strategy that they used.

Funding

R.D. acknowledges support from a Sir Henry Dale Fellowship jointly funded by the Wellcome Trust and the Royal Society (Grant Number 107617/Z/15/Z). The Centre for Public Health Research is supported by a Programme Grant from the Health Research Council of New Zealand. The research leading to these results has received funding from the European Research Council under the European Union’s Seventh Framework Programme (FP7/2007-2013) / ERC grant agreement no 668954. Click here for additional data file.
  29 in total

1.  When should epidemiologic regressions use random coefficients?

Authors:  S Greenland
Journal:  Biometrics       Date:  2000-09       Impact factor: 2.571

2.  Effects of adjusting for instrumental variables on bias and precision of effect estimates.

Authors:  Jessica A Myers; Jeremy A Rassen; Joshua J Gagne; Krista F Huybrechts; Sebastian Schneeweiss; Kenneth J Rothman; Marshall M Joffe; Robert J Glynn
Journal:  Am J Epidemiol       Date:  2011-10-24       Impact factor: 4.897

Review 3.  Bayesian perspectives for epidemiological research. II. Regression analysis.

Authors:  Sander Greenland
Journal:  Int J Epidemiol       Date:  2007-02-28       Impact factor: 7.196

4.  Methods of covariate selection: directed acyclic graphs and the change-in-estimate procedure.

Authors:  Hsin-Yi Weng; Ya-Hui Hsueh; Locksley L McV Messam; Irva Hertz-Picciotto
Journal:  Am J Epidemiol       Date:  2009-04-10       Impact factor: 4.897

5.  Exhaustion, automation, theory, and confounding.

Authors:  Marshall M Joffe
Journal:  Epidemiology       Date:  2009-07       Impact factor: 4.822

6.  The impact of confounder selection criteria on effect estimation.

Authors:  R M Mickey; S Greenland
Journal:  Am J Epidemiol       Date:  1989-01       Impact factor: 4.897

7.  The interpretation of multiplicative-model parameters as standardized parameters.

Authors:  S Greenland; G Maldonado
Journal:  Stat Med       Date:  1994-05-30       Impact factor: 2.373

8.  Atopy and asthma in rural Poland: a paradigm for the emergence of childhood respiratory allergies in Europe.

Authors:  B Sozanska; S J Macneill; M Kajderowicz-Kowalik; H Danielewicz; M Wheatley; A J Newman Taylor; A Boznanski; P Cullinan
Journal:  Allergy       Date:  2007-04       Impact factor: 13.146

9.  On the definition of a confounder.

Authors:  Tyler J VanderWeele; Ilya Shpitser
Journal:  Ann Stat       Date:  2013-02       Impact factor: 4.028

10.  Sodium intake in infancy and blood pressure at 7 years: findings from the Avon Longitudinal Study of Parents and Children.

Authors:  M-J Brion; A R Ness; G Davey Smith; P Emmett; I Rogers; P Whincup; D A Lawlor
Journal:  Eur J Clin Nutr       Date:  2007-07-11       Impact factor: 4.016

View more
  71 in total

1.  Psychosocial risk factors and asthma among adults in Puerto Rico.

Authors:  Yueh-Ying Han; Erick Forno; Glorisa Canino; Juan C Celedón
Journal:  J Asthma       Date:  2018-06-06       Impact factor: 2.515

2.  Identifying patients at higher risk of pneumonia after lung resection.

Authors:  Gan-Wei Liu; Xi-Zhao Sui; Shao-Dong Wang; Hui Zhao; Jun Wang
Journal:  J Thorac Dis       Date:  2017-05       Impact factor: 2.895

3.  Visualization tool of variable selection in bias-variance tradeoff for inverse probability weights.

Authors:  Ya-Hui Yu; Kristian B Filion; Lisa M Bodnar; Maria M Brooks; Robert W Platt; Katherine P Himes; Ashley I Naimi
Journal:  Ann Epidemiol       Date:  2019-12-13       Impact factor: 3.797

4.  Ethnic Minority Status and Experiences of Infertility in Female Veterans.

Authors:  Rachel P Goossen; Karen M Summers; Ginny L Ryan; Michelle A Mengeling; Brenda M Booth; James C Torner; Craig H Syrop; Anne G Sadler
Journal:  J Womens Health (Larchmt)       Date:  2018-11-27       Impact factor: 2.681

5.  Chronic hypnotic use at 10 years-does the brand matter?

Authors:  Yochai Schonmann; Or Goren; Ronen Bareket; Doron Comaneshter; Arnon D Cohen; Shlomo Vinker
Journal:  Eur J Clin Pharmacol       Date:  2018-08-08       Impact factor: 2.953

6.  Caution: work in progress : While the methodological "revolution" deserves in-depth study, clinical researchers and senior epidemiologists should not be disenfranchised.

Authors:  Miquel Porta; Francisco Bolúmar
Journal:  Eur J Epidemiol       Date:  2016-07-14       Impact factor: 8.082

7.  Selecting Shrinkage Parameters for Effect Estimation: The Multi-Ethnic Study of Atherosclerosis.

Authors:  Joshua P Keller; Kenneth M Rice
Journal:  Am J Epidemiol       Date:  2018-02-01       Impact factor: 4.897

8.  Influenza vaccine effectiveness among patients with high-risk medical conditions in the United States, 2012-2016.

Authors:  Mei Shang; Jessie R Chung; Michael L Jackson; Lisa A Jackson; Arnold S Monto; Emily T Martin; Edward A Belongia; Huong Q McLean; Manjusha Gaglani; Kempapura Murthy; Richard K Zimmerman; Mary Patricia Nowalk; Alicia M Fry; Brendan Flannery
Journal:  Vaccine       Date:  2018-11-09       Impact factor: 3.641

9.  Influence of Modeling Choices on Value of Information Analysis: An Empirical Analysis from a Real-World Experiment.

Authors:  David D Kim; Gregory F Guzauskas; Caroline S Bennette; Anirban Basu; David L Veenstra; Scott D Ramsey; Josh J Carlson
Journal:  Pharmacoeconomics       Date:  2020-02       Impact factor: 4.981

10.  First-year weight loss with androgen-deprivation therapy increases risks of prostate cancer progression and prostate cancer-specific mortality: results from SEARCH.

Authors:  Kagan Griffin; Ilona Csizmadi; Lauren E Howard; Gina-Maria Pomann; William J Aronson; Christopher J Kane; Christopher L Amling; Matthew R Cooperberg; Martha K Terris; Jennifer Beebe-Dimmer; Stephen J Freedland
Journal:  Cancer Causes Control       Date:  2019-01-30       Impact factor: 2.506

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.