Literature DB >> 28068910

Outcome-sensitive multiple imputation: a simulation study.

Evangelos Kontopantelis1,2, Ian R White3, Matthew Sperrin4, Iain Buchan4.   

Abstract

BACKGROUND: Multiple imputation is frequently used to deal with missing data in healthcare research. Although it is known that the outcome should be included in the imputation model when imputing missing covariate values, it is not known whether it should be imputed. Similarly no clear recommendations exist on: the utility of incorporating a secondary outcome, if available, in the imputation model; the level of protection offered when data are missing not-at-random; the implications of the dataset size and missingness levels.
METHODS: We used realistic assumptions to generate thousands of datasets across a broad spectrum of contexts: three mechanisms of missingness (completely at random; at random; not at random); varying extents of missingness (20-80% missing data); and different sample sizes (1,000 or 10,000 cases). For each context we quantified the performance of a complete case analysis and seven multiple imputation methods which deleted cases with missing outcome before imputation, after imputation or not at all; included or did not include the outcome in the imputation models; and included or did not include a secondary outcome in the imputation models. Methods were compared on mean absolute error, bias, coverage and power over 1,000 datasets for each scenario.
RESULTS: Overall, there was very little to separate multiple imputation methods which included the outcome in the imputation model. Even when missingness was quite extensive, all multiple imputation approaches performed well. Incorporating a secondary outcome, moderately correlated with the outcome of interest, made very little difference. The dataset size and the extent of missingness affected performance, as expected. Multiple imputation methods protected less well against missingness not at random, but did offer some protection.
CONCLUSIONS: As long as the outcome is included in the imputation model, there are very small performance differences between the possible multiple imputation approaches: no outcome imputation, imputation or imputation and deletion. All informative covariates, even with very high levels of missingness, should be included in the multiple imputation model. Multiple imputation offers some protection against a simple missing not at random mechanism.

Entities:  

Keywords:  Imputed outcome; Missing data; Missingness; Multiple imputation

Mesh:

Year:  2017        PMID: 28068910      PMCID: PMC5220613          DOI: 10.1186/s12874-016-0281-5

Source DB:  PubMed          Journal:  BMC Med Res Methodol        ISSN: 1471-2288            Impact factor:   4.615


Background

Missing data is a common obstacle to observational health science and its pitfalls are well known [1]. To exclude study subjects with any missing covariate observations, so called complete case analysis, is at best of low statistical power and at worst provides biased estimates. The complexity of the missing data problem, or obtaining accurate inferential estimates in the presence of missing data, depends on the nature of the mechanism by which data are missing [2]. The less problematic scenario occurs when the probability of an observable data point being missing (the missingness probability) does not depend on any observed or unobserved parameters, and this missingness mechanism is known as Missing Completely at Random (MCAR). However, MCAR mechanisms are considered rare in practice, especially for surveys [3]. More commonly, the missingness probability depends on observed variables, and hence it can be accounted for by the information contained in the dataset. This missing data mechanism has been labelled Missing at Random (MAR). Finally, the most challenging missingness mechanism occurs when the missingness probability depends on unobserved values, and called Missing Not at Random (MNAR). Using self-reporting of sexual activity we can explore MAR and MNAR examples. A MAR scenario would arise if girls are less likely than boys to report whether they are sexually active, but sexually active teenagers are no more likely to report than those non-active. However, if sexually active girls are less likely to report than non-active girls, this would be an MNAR scenario. Although suboptimal approaches to imputation are still routinely used [4], multiple imputation has been accepted by methodologists as the most appropriate framework for dealing with MCAR and MAR mechanisms [5]. Multiple imputation can be described in three steps: (a) drawing the missing data from their posterior predictive distribution under a posited Bayesian model, across N datasets; (b) analysing each dataset separately with a chosen method, usually a regression model; and (c) pooling the estimates and their standard errors across the N analyses using Rubin’s rules [2], allowing for the use of the within-imputation and between-estimation variation components in the calculations. Multiple imputation has a largely Bayesian rationale but it also works well in frequentist applications by providing nominal coverage levels and unbiased point estimates [2]. When the analysis model (Step 2) is Bayesian the resulting framework is fully Bayesian; alternatively, frequentist maximum-likelihood estimates can be used in the model to draw missing values [6, 7]. An interesting trait of multiple imputation is that it performs better at imputing missing predictors when outcome information is included in the models [8]. Although standard applications of multiple imputation do not deal with MNAR mechanisms, they can offer some protection against them [9]. Furthermore, multiple imputation can accommodate MNAR scenarios flexibly and is thus well-suited to sensitivity analyses [10]. This paper tackles five outstanding issues about multiple imputation. von Hippel argued that researchers should impute values for the outcome, but exclude cases with imputed outcomes when fitting the substantive model [11]. It is unclear when this is the best strategy compared with not imputing the outcome on the one hand, and using imputed outcomes in the analysis on the other hand. Another question of practical interest is how much missingness should be considered manageable within a multiple imputation framework, i.e. is the performance of multiple imputation consistent as missingness increases and is there a level above which performance deteriorates to such an extent that it makes the method and data of little practical use? Although this has been answered for multiple imputation that does not impute the outcome, which was found to perform consistently across all missingness levels [12], the question remains for outcome imputations. Also unknown is the role of multiple correlated outcomes in the imputation models. It is not uncommon for studies to collect information on two or more correlated outcomes. However, these outcomes are often analysed through separate multiple imputation models, which do not utilise the association of the outcome of interest with a second (or more) available outcome. Would the inclusion of a second outcome lead to an improved multiple imputation model? In addition, as far as we know, the level of protection, if any, offered by current multiple imputation methods (i.e. methods that assume MAR) against MNAR mechanisms has not been quantified within a simulations framework. Although there are technical challenges and numerous assumptions when developing such a framework, a high level of protection would make researchers more confident when reporting results from analyses with missing data—especially since MNAR mechanisms cannot be identified without additional external data or prior knowledge. Finally, the size of the investigated dataset could be an important parameter and the performance of multiple imputation has largely been assessed in small or moderate datasets [8, 13], mainly for computational reasons. We have chosen scenarios of 1000 cases, which would be relevant to small and moderate studies, and more or less in agreement with previous investigations. However, it is now common to analyse data from many thousands or even millions of people, for example using Electronic Health Records (EHRs), and for this reason we also analysed scenarios of 10,000 cases. In this paper we address all of the questions above using simulations. We simulate a wide range of scenarios which are not uncommon in observational studies with databases of routinely collected data, where hundreds of variables may be available and often have varying levels of missingness [14]. Although our motivation stems from our experiences with observational data, our findings are also relevant to clinical trials data, which are usually less variable (e.g. levels of missingness across variables are uniform, when someone is lost to follow-up).

Methods

Obtaining algebraic answers to the questions we have posed is challenging due to: the large number of parameters often involved; and the asymptotic estimation approaches commonly employed. Arguably the most reasonable approach is to use realistically simulated data, where the true associations between predictors and the outcome are known, and can be used reliably to quantify method performance. The processes described below were repeated 1,000 times, to obtain different datasets under the specified parameters and to analyse them.

Data generation

We assumed two dataset sizes of 1,000 and 10,000 patients for which we originally had complete information on a primary binary outcome Y, a secondary binary outcome Y’, a binary exposure variable E and a continuous covariate X confounding the relationship between exposure and outcomes. The whole process was implemented in Stata v14.1 [15], and the code is provided in Online Additional file 1. We used the drawnorm command to draw observations from multivariate normal distributions, allowing for a ≈ 0.4 Pearson’s correlation between X and E. Covariate X had mean 0 and variance 1, while we set Pr (E = 1) = 0.5. Outcome Y’ was generated last, correlated with primary outcome Y (tetrachoric ρ ≈ 0.49), but independent of X and E given Y. Writing π = p(Y = 1|X, E), we assumed the logistic regression model logit(π) = β 0 + β 1 E + β 2 X + β 3 E ⋅ X with parameters: exp (β 0) ≈ 0.091, exp (β 1) = 2, exp (β 2) = 1.5 and exp (β 3) = 1.2. These parameters lead to conditional probabilities for the outcome of Pr (Y = 1|E = 0) = 0.091 and Pr (Y = 1|E = 1) = 0.232. Weaker associations exist between Y’ and E, Y’ and X, Y ’ and X * E, but are not of interest. The data structure is displayed in Fig. 1.
Fig. 1

Data structure

Data structure

Missingness mechanisms

We implemented three missing data mechanisms (MCAR, MAR and MNAR) and four levels of overall missingness for each variable (20%, 40%, 60% and 80%). In each case, covariate X and outcomes Y and Y ’ all had the same level of missingness but information for exposure E was always complete. In the MCAR setting, values for X, Y and Y’ were independently set to be missing. In the MAR setting, the probability to be missing for each of X, Y or Y’ was independently set to be conditional on the exposure E with OR = 5. In other words, the odds of a missing value for X, Y or Y’ were five times as high in the presence of the exposure (i.e. when E = 1). In the final setting, a relatively simple MNAR scenario, the probabilities of missing data for X, Y or Y’ were conditional on the true values of X, Y or Y’ respectively, with OR = 5 used across each of the associations. Information for exposure E was always complete. However, the MNAR scenario is of course not exhaustive and alternative MNAR mechanisms could vary across exposure groups [16]. It should also be noted that although the missingness mechanism we modelled is rather extreme, it was a conscious decision to make it more likely to observe performance variability across models. We anticipate our findings to be relevant to weaker associations, where model performances are expected be less variable.

Alternative data structures

We also considered two alternative structures, as sensitivity analyses, which we do not present in detail in this paper but the code for which is available from the authors. In the first sensitivity analysis we simulated a continuous rather than a binary outcome, and in the second sensitivity analysis we included a second covariate X’ to which we applied the same missingness mechanisms.

Analysis

Across each missingness mechanism we followed the same seven logistic regression analyses, seven of which were multiple imputation approaches (Table 1). The first analysis (A) was the simplest, a complete cases analysis, with the sole purpose of providing a benchmark against which to compare the multiple imputation approaches. In the remaining seven analyses we used the mi family of commands in Stata, with mi impute chained for the imputation and mi estimate with a multiple logistic regression (logit command) for the analysis. Analyses B, C, D and E all ignored the secondary outcome Y’. In the second analysis (B), which is known to give biased estimates [11], we excluded cases where the outcome was missing and the outcome was not included in the imputation model. In the third analysis (C), we again dropped missing outcome cases but included the outcome in the imputation model for the missing covariate X. In the fourth analysis (D) we included the outcome in the imputation model and imputed its missing values as well as missing values for X. An alternative approach suggested by von Hippel [11] was our analysis E, which followed D and included the outcome in the multiple imputation model and imputing it, but deleted cases where the outcome was imputed. Analyses F and G, followed C and D respectively but also included the second outcome in the imputation models and imputed their missing values. Finally, analysis H followed the setup of D, except it did not include the covariate X in the multiple imputation or analysis models. This aimed to assess whether the covariate should be included, even when its missingness levels were very high.
Table 1

Analysis methods

Acomplete case analysis (no multiple imputation [mi])
Bno outcome imputation, not included in mi model
Ca no outcome imputation, outcome imputed in mi model
Da outcome imputed and included in mi model
Ea outcome imputed and included in mi model but then cases where it was imputed are deleted
Fa as in C but also including a second correlated outcome in the mi model
Ga as in D but also including a second correlated outcome in the mi model
Has in D but the mi and analysis models do not include the covariate

Main models of interest, other models provided for comparison purposes

Analysis methods Main models of interest, other models provided for comparison purposes The analysis approaches and the data setup were selected to fit our research questions. Comparing analysis models C and D, which are the commonly recommended best practice models, will provide information on whether imputing an incomplete outcome is preferable to excluding the relevant cases. Model E, which deletes cases where the outcome is imputed, will be assessed as a best practice alternative to C and D. Comparisons between C and F, as well as D and G will inform us whether the inclusion of a second outcome which is correlated to our outcome of interest leads to a better multiple imputation model. Comparing models that have included the covariate (e.g. C and D) with model H, across various levels of missingness, will answer whether it is preferable to exclude a covariate from a multiple imputation model when most of its values are missing. Simulating different missingness mechanisms will allow us to quantify the performance of multiple imputation approaches vs complete case analyses in the most problematic MNAR scenario, compared to the known protection it offers for the most common MCAR and MCAR scenarios [13]. Finally, repeating the analyses in datasets of different sizes will shed light to whether our conclusions are sample-size dependent or not.

Performance measures

We aimed to measure the performance of all multiple imputation and analysis approaches with logistic regression, across the scenarios of missingness described above and over 1,000 iterations, in the estimation of the three true association of interest: E →  Y (our main focus), X →  Y and X * E →  Y. There are numerous performance measures that can be used in simulation studies [17], but we considered mean absolute error, mean bias, coverage probability and power of the analyses in relation to the three parameters of interest to be adequate for our investigation. Mean absolute error was calculated as where z is one of the three parameters of interest, expressed as a log-odds ratio. Analogously mean bias is the mean difference in the estimate to the true parameter, and was calculated as . Therefore, assuming the used exposure effect of log[2], a reported mean bias of 0.1 or −0.1 would mean that the returned estimate was on average ≈ log(1.81) or ≈ log(2.21), respectively. The coverage probability, is the proportion of 95% confidence intervals for the estimate that contain the true parameter across the 1,000 iterations. Theoretically this should be close to 95% but the bias introduced through the MAR and MNAR mechanisms can affect coverage levels. Finally, we calculated the power to detect that the parameter is different from zero by computing the proportion of the 1,000 95% confidence intervals for each parameter that did not include zero. However, power needs to be carefully interpreted in the presence of bias since bias will move the estimate closer or further away from the alternative hypothesis on which power is calculated, and in the latter case higher bias will lead to higher power. Nevertheless, provided bias is similar across the methods to compare, power can be used for comparisons, even if bias is not zero, and we felt it was an important metric that would complement the study.

Results

Results for the coefficient of the exposure β 1 are presented in Table 2 and Table 3 for datasets of 1,000 and 10,000 cases, respectively. Figures 2, 3, 4 and 5 present the exposure performance metrics with their respective error bars. Although all methods successfully converged for the larger datasets, there was some variation in the smaller datasets for very high levels of missingness (Online Additional file 2: Table S5). Results for the coefficients of the covariate and the exposure-covariate interaction are also presented in Online Additional file 2: Table S1, S2, S3 and S4).
Table 2

Performance results for exposure E, datasets of 1,000 observationsa

% missABCb Db Eb Fb Gb H
MCARMean biasc 20−0.019−0.107−0.020−0.025−0.030−0.021−0.023−0.427
40−0.038−0.209−0.041−0.045−0.020−0.041−0.045−0.437
60−0.084−0.301−0.064−0.056−0.060−0.067−0.055−0.440
800.134−0.409−0.109−0.092−0.145−0.126−0.097−0.458
Mean errorc 200.2180.2080.2000.2010.1990.1990.1990.430
400.2900.2680.2320.2310.2290.2320.2330.444
600.4750.3610.3060.3130.3080.3090.3010.466
801.0160.5030.5840.5030.5430.5600.4890.525
Coverage200.9500.9410.9430.9450.9490.9470.9430.489
400.9620.9130.9630.9570.9590.9620.9490.605
600.9650.9070.9590.9390.9550.9600.9430.718
800.9920.9570.9890.9560.9860.9880.9650.823
Power200.7200.6880.7710.7610.7770.7690.7700.274
400.4620.4250.6040.5930.6440.6100.5970.181
600.2280.2130.3670.4130.3940.3730.4330.163
800.0650.0620.0680.1800.0980.0740.1690.124
MARMean biasc 20−0.019−0.125−0.030−0.035−0.030−0.030−0.035−0.425
40−0.014−0.227−0.046−0.065−0.044−0.050−0.062−0.441
60−0.046−0.308−0.063−0.064−0.049−0.059−0.062−0.446
80−0.054−0.350−0.069−0.119−0.080−0.073−0.128−0.408
Mean errorc 200.2240.2150.2000.2000.2030.1980.1970.428
400.2900.2820.2280.2270.2340.2290.2290.448
600.4910.3670.3140.3120.3030.3140.3050.467
801.0700.4630.5930.5010.5460.6010.5020.494
Coverage200.9410.9250.9550.9540.9560.9550.9530.504
400.9580.8960.9530.9560.9500.9580.9480.561
600.9550.8890.9630.9480.9710.9640.9460.706
800.9940.9430.9780.9540.9780.9780.9460.836
Power200.7080.6780.7790.7630.7710.7740.7790.271
400.4480.3720.5830.5640.6060.5870.5930.193
600.2240.1950.3500.3910.3730.3600.4040.161
800.0100.0520.0500.1470.0770.0450.1430.123
MNARMean biasc 20−0.026−0.150−0.047−0.044−0.054−0.047−0.046−0.425
40−0.092−0.302−0.135−0.113−0.097−0.139−0.121−0.454
60−0.086−0.334−0.050−0.022−0.027−0.052−0.021−0.450
800.038−0.478−0.253−0.283−0.314−0.316−0.275−0.484
Mean errorc 200.2270.2280.2070.2070.2090.2080.2070.431
400.3750.3710.3020.2930.2920.3040.2930.472
600.5900.4110.3660.3680.3610.3710.3550.479
801.2830.7410.8410.7370.7730.8810.7110.654
Coverage200.9540.9230.9400.9450.9410.9440.9430.554
400.9570.9270.9580.9500.9610.9540.9490.708
600.9730.9620.9670.9510.9640.9680.9460.746
800.9971.0000.9930.9900.9990.9950.9860.921
Power200.6520.5920.7140.7120.6940.7060.7130.237
400.3070.2230.3830.3980.4200.3890.3990.153
600.2090.1480.3300.3680.3660.3270.3820.152
800.0150.0100.0410.0880.0390.0300.0960.092

aAnalysis model A: complete case analysis (no multiple imputation [mi]); B: no outcome imputation, not included in mi model; C: no outcome imputation, outcome imputed in mi model; D: outcome imputed and included in mi model; E: outcome imputed and included in mi model but then observations where it was imputed are deleted; F as in C but also including a second correlated outcome in the mi model; G as in D but also including a second correlated outcome in the mi model; H as in D but the mi and analysis models do not include the covariate

bMain models of interest, other models provided for comparison purposes

cReported on log-odds scale and based on a true effect of log [2]

Table 3

Performance results for exposure E, datasets of 10,000 observationsa

% missABCb Db Eb Fb Gb H
MCARMean biasc 200.001−0.089−0.003−0.008−0.006−0.003−0.006−0.413
40−0.005−0.181−0.016−0.024−0.009−0.016−0.021−0.419
60−0.017−0.268−0.027−0.035−0.017−0.028−0.032−0.421
80−0.021−0.343−0.023−0.036−0.033−0.027−0.032−0.418
Mean errorc 200.0640.0940.0590.0590.0600.0590.0590.413
400.0880.1820.0710.0720.0710.0700.0710.419
600.1350.2680.0940.0940.0880.0940.0950.421
800.2700.3430.1500.1440.1480.1480.1430.418
Coverage200.9620.7920.9590.9650.9480.9590.9590.000
400.9460.4390.9540.9460.9580.9540.9520.000
600.9430.2910.9560.9310.9650.9580.9390.007
800.9520.3940.9520.9290.9500.9490.9290.137
Power201.0001.0001.0001.0001.0001.0001.0000.983
401.0001.0001.0001.0001.0001.0001.0000.935
600.9740.9821.0001.0001.0001.0001.0000.772
800.5180.6350.9490.9430.9560.9460.9440.547
MARMean biasc 200.005−0.101−0.005−0.012−0.009−0.005−0.013−0.412
400.001−0.204−0.031−0.043−0.032−0.031−0.042−0.419
60−0.015−0.275−0.038−0.046−0.028−0.037−0.046−0.423
800.014−0.349−0.058−0.068−0.058−0.058−0.072−0.415
Mean errorc 200.0640.1040.0590.0580.0600.0580.0580.412
400.0870.2040.0720.0750.0740.0720.0750.419
600.1350.2750.0960.0980.0920.0950.0970.423
800.3040.3490.1600.1590.1530.1580.1610.415
Coverage200.9650.7270.9630.9670.9620.9650.9640.000
400.9520.3340.9580.9400.9470.9510.9350.000
600.9480.2370.9500.9240.9530.9500.9180.004
800.9440.3560.9350.9180.9430.9410.9260.161
Power201.0001.0001.0001.0001.0001.0001.0000.987
401.0001.0001.0001.0001.0001.0001.0000.944
600.9770.9820.9991.0001.0001.0001.0000.792
800.4860.6130.9050.8970.9200.9120.9020.537
MNARMean biasc 200.003−0.125−0.023−0.021−0.024−0.023−0.024−0.411
40−0.006−0.250−0.091−0.072−0.080−0.091−0.081−0.417
60−0.003−0.288−0.0160.0100.005−0.0170.003−0.420
80−0.026−0.358−0.186−0.161−0.186−0.182−0.176−0.427
Mean errorc 200.0670.1280.0630.0630.0660.0630.0630.411
400.1120.2500.1130.1020.1070.1130.1070.417
600.1500.2880.1030.1050.1060.1040.1030.420
800.4560.3670.2530.2370.2580.2520.2410.428
Coverage200.9520.6430.9480.9470.9350.9490.9470.000
400.9600.3490.8930.9080.8860.8870.8930.002
600.9520.3940.9550.9440.9370.9610.9470.017
800.9670.9430.9810.9160.9710.9810.9160.360
Power201.0001.0001.0001.0001.0001.0001.0000.973
400.9980.9851.0001.0001.0001.0001.0000.826
600.9420.8980.9990.9971.0000.9960.9990.738
800.2560.0940.3580.5270.3940.3640.5190.331

aAnalysis model A: complete case analysis (no multiple imputation [mi]); B: no outcome imputation, not included in mi model; C: no outcome imputation, outcome imputed in mi model; D: outcome imputed and included in mi model; E: outcome imputed and included in mi model but then observations where it was imputed are deleted; F as in C but also including a second correlated outcome in the mi model; G as in D but also including a second correlated outcome in the mi model; H as in D but the mi and analysis models do not include the covariate

bMain models of interest, other models provided for comparison purposes

cReported on log-odds scale and based on a true effect of log [2]

Fig. 2

Mean Bias and 95% Confidence Intervals for exposure E in datasets of 1000 (top) and 10,000 observations (bottom)

Fig. 3

Mean absolute Error and 95% Confidence Intervals for exposure E in datasets of 1000 (top) and 10,000 observations (bottom)

Fig. 4

Coverage and 95% Confidence Intervals for exposure E in datasets of 1000 (top) and 10,000 observations (bottom)

Fig. 5

Power and 95% Confidence Intervals for exposure E in datasets of 1000 (top) and 10,000 observations (bottom)

Performance results for exposure E, datasets of 1,000 observationsa aAnalysis model A: complete case analysis (no multiple imputation [mi]); B: no outcome imputation, not included in mi model; C: no outcome imputation, outcome imputed in mi model; D: outcome imputed and included in mi model; E: outcome imputed and included in mi model but then observations where it was imputed are deleted; F as in C but also including a second correlated outcome in the mi model; G as in D but also including a second correlated outcome in the mi model; H as in D but the mi and analysis models do not include the covariate bMain models of interest, other models provided for comparison purposes cReported on log-odds scale and based on a true effect of log [2] Performance results for exposure E, datasets of 10,000 observationsa aAnalysis model A: complete case analysis (no multiple imputation [mi]); B: no outcome imputation, not included in mi model; C: no outcome imputation, outcome imputed in mi model; D: outcome imputed and included in mi model; E: outcome imputed and included in mi model but then observations where it was imputed are deleted; F as in C but also including a second correlated outcome in the mi model; G as in D but also including a second correlated outcome in the mi model; H as in D but the mi and analysis models do not include the covariate bMain models of interest, other models provided for comparison purposes cReported on log-odds scale and based on a true effect of log [2] Mean Bias and 95% Confidence Intervals for exposure E in datasets of 1000 (top) and 10,000 observations (bottom) Mean absolute Error and 95% Confidence Intervals for exposure E in datasets of 1000 (top) and 10,000 observations (bottom) Coverage and 95% Confidence Intervals for exposure E in datasets of 1000 (top) and 10,000 observations (bottom) Power and 95% Confidence Intervals for exposure E in datasets of 1000 (top) and 10,000 observations (bottom)

Mean bias

In smaller datasets and for MCAR and MAR data, levels of bias were low across most models, except in B (outcome not included in the imputation model) and H (covariate not included in the multiple imputation model). Complete case analysis (model A) was often the best performer, especially for low levels of missingness but results could only be obtained for a subsample of less problematic datasets, due to perfect prediction or non-convergence (Online Additional file 2: Table S5). Bias levels increased for MNAR data. In larger datasets bias levels were lower for all methods and complete case analysis appeared to be by the best performer with very low bias in every simulation scenario.

Mean absolute error

In both smaller and larger datasets, the best performing models were: C (no outcome imputation, outcome included in mi model); D (outcome imputed and included in mi model); E (outcome imputed and included in mi model and then deleted); F (no outcome imputation, both outcomes included in mi model); and G (both outcomes included in mi model, outcome of interest imputed). For the smaller datasets, the models that imputed the outcome (D, E and G) generally performed only slightly better for very high levels of missingness in MCAR data, and for MAR data (and especially for higher rates of missingness). Error increased with increasing missingness but was not too dissimilar across the three missingness mechanisms. In the larger datasets, levels of mean absolute error were much lower and there was no benefit in using the second outcome, with models C, D and E performing the best, with variations in different settings. Overall, the best model was E but only slightly better than C and D.

Coverage

Again, at both dataset sizes, models C, D, E, F and G performed best. There was very little to separate them, however, the models that imputed the outcome (D and G) tended to be closer to the nominal 95%. There were relatively small differences across missingness mechanisms and coverage levels were good in all scenarios, with the lowest rates amongst the five top performing models observed for D and G in MAR data and high missingness levels. Similarly, coverage rates were consistently high across all levels of missingness. In larger datasets, there was no benefit to using a second outcome, with models C, D and E equivalent in almost all missingness scenarios. Overall, differences between models C, D and E are small but E had better coverage in the larger datasets for extensive missingness.

Power

Results for power were consistent with error and coverage, with models C, D, E, F and G again performing best. In the smaller datasets, there were small differences between these models, except for very high levels of missingness, and especially for MCAR and MAR mechanisms, where imputing the outcome (models D and G) returned higher power level, albeit still very low. However, for lower levels of missingness, model C performed well and, more often than not, slightly better than D. The nature of the missingness mechanism had some effect on power, with lower levels observed for MNAR data, especially as levels of missingness increased. As expected, the more data are missing the lower the power, and all models performed very poorly for high or very high levels of missingness (60% or above). In larger datasets, the picture did not change with models C, D, E, F and G being almost equivalent, but model D performed better for extensive missingness (60% or above) in MNAR data. Overall, E outperformed C is all settings and D for low and moderate levels of missingness, while D performed better for very high levels of missingness.

Sensitivity analyses

Patterns of results in the two sets of sensitivity simulations broadly agreed with what we observed for the main simulations and further supported our findings. When analysing a continuous outcome, differences between multiple imputation models were again very small. Focusing on datasets of 1000 observations and the multiple imputation models of main interest (C to G), mean bias was very similar for all missingness mechanisms and levels. Mean bias was very close to zero for all MCAR and MAR settings, except for 80% levels of missingness. For MNAR data, mean bias was very close to zero for 20% missingness and linearly increased with missingness. Mean absolute error was again similar in these methods across all missingness mechanisms, with some variability being observed for 80% missingness and method G (outcome imputed and included in mi model, including secondary outcome) performing slightly better in those scenarios (Online Additional file 2: Figure S1). Coverage was similar in all scenarios, except for high levels of missingness where the outcome imputation models (G and especially D) slightly underperformed. However, that shortcoming was counterbalanced for model G by higher power in all scenarios except very low levels of missingness, where there was very little variation in performance (Online Additional file 2: Figure S2).

Discussion

Our results indicate that in general, there are very small differences between models that impute the outcome compared with those that do not, when all else is equal and the outcome is included in the imputation model. However, in some contexts small differences emerge that should underpin recommendations as to the choice of model. The von Hippel approach [11], our model E, where the outcome is included in the imputation model and imputed but cases where the outcome is imputed are later dropped performed well. However, the differences between this approach and alternative models, where the outcome is not imputed or imputed and not dropped, were generally very small if any (error bars for all performance metrics overlapped substantially). Furthermore, the von Hippel approach was not consistently better in all scenarios. Another consideration is the presence of an “auxiliary” variable, a variable that in not part of the analysis model but is used in the multiple imputation to improve the prediction of missing values. If such a variable is associated with missingness in the outcome, model E is known to produce biased parameter estimates and should be avoided [18]. The level of missingness naturally affects the performance of the multiple imputation models, especially with regards to power (primarily) and error (secondarily). However, in agreement with Janssen et al. [12], we recommend using all available data even when missingness among covariates of interest is extensive. Multiple imputation models that exclude such covariates seem to perform much worse. For very high levels of missingness and moderately sized datasets we recommend the use of simulation-based platforms to estimate the power to detect effects [19]. Convergence was not an issue with any models when the datasets contained 10,000 observations, but it was a factor to consider in the 1,000 observations datasets as the level of missingness increased. Multiple imputation models that did not impute the outcome and were only modestly affected, while complete case analysis was severely affected. The size of the datasets (1,000 or 10,000) did not substantially affect how the models ranked within each group. Interestingly, in the larger datasets, a complete case analysis approach was generally only slightly worse than the best performing multiple imputation models for low levels of missingness. Therefore, existing multiple imputation approaches may be less relevant to large health informatics databases than to randomised clinical trials. Surprisingly the inclusion of a second outcome in the multiple imputation model, moderately correlated to the primary outcome, made very little difference to performance. Since for the imputation model there is no real distinction between predictors and outcomes, we would expect the inclusion of the secondary outcome to lead to improved performance. However, our findings could be explained by the associations between the predictors and the secondary outcome. In other words, the secondary outcome has little independent information to add to the model. A weaker association between predictors and secondary outcomes and a stronger correlation between outcomes would make the secondary outcome a useful addition to the multiple imputation model. However, we did observe slightly better performance for the model in the continuous outcome sensitivity analysis, for some scenarios, mainly in terms of power but also mean absolute error. Hence a more complete multiple imputation model that includes all outcomes is recommended. Finally, although all models performed worse when data were MNAR, multiple imputation models can offer some protection, in terms of mean absolute error, even in this relatively extreme missingness scenario we simulated (OR = 5 for the missingness mechanism). Multiple imputation models outperformed complete case analyses in both smaller and larger datasets. However, the benefits of using multiple imputation methods were not as high for MNAR as for MCAR or MAR data, and were more obvious in the smaller datasets.

Strengths and limitations

We have evaluated the performance of commonly used imputation approaches in realistic simulated data scenarios. Nevertheless, some limitations exist. First, although realistic, our simulated scenarios cannot be exhaustive and results may vary in alternative scenarios with different hypothesised associations between exposure, covariate and outcome and different distributions. However, we would expect the methods to perform similarly, at least relatively to each other, and our conclusions not to be affected—at least in MCAR settings. Our MAR settings made complete cases analysis (method A) unbiased because missingness depended only on exposure; if missingness of covariates had also depended on outcome then bias would have arisen in complete cases analysis. Regarding MNAR, we investigated common scenarios but there are many other possible mechanisms and our findings are not generalisable to them. In particular, our MNAR mechanism for Y was akin to case–control sampling and hence caused no bias. A different missing data mechanism that depended on both Y and E could cause large bias in the coefficient of E, especially if the association between missingness and Y differed across exposure groups [16]. Second, the precision obtained with simulations of 1,000 iterations is not ideal but the models we executed are complex and require considerable computational time. Third, a sample of 1,000 might seem too large if compared against trial data, but it was a necessity if we were to investigate very highc rates of missingness. Fourth, the substantive model was not entirely consistent with the imputation model because of the interaction term –we felt it was important to reflect this approach because it is often seen in practice. Fifth, we only considered one strength of association between the outcome Y and the secondary outcome Y’: although we modelled a rather strong association, probably stronger to what would be observed in practice in most cases, even stronger associations are likely to increase the value of including the secondary outcome in imputation models. Finally, the computational time led us to select our largest simulated dataset to include 10,000 cases. Unfortunately, this is not necessarily representative of a contemporary electronic health records dataset which can hold hundreds of thousands or millions of cases. However, even that limited size is very different to the size of a clinical trial, on which multiple imputation methods have been routinely evaluated in the past. Therefore, we argue that we manage to provide an incomplete view on the relevance of these methods in larger datasets.

Conclusions

There was very little to separate the multiple imputation methods of interest. Although the method that imputes the outcome of interest and then removes observations where the outcome is imputed performed slightly better in some scenarios, especially for low and moderate levels of missingness, it was not always better and it is known to be biased in the presence of auxiliary variables. For very high levels of missingness, the higher power obtained when imputing the outcome (and not dropping observations) might make this approach somewhat more appealing. However, as long as the outcome is included in the imputation model, the choice of the multiple imputation approach makes no practical difference. Important covariates need to be included in the imputation models even when their levels of missingness are very high. Although the use of secondary outcomes did not lead to substantially better models in our simulations, some improvements were observed in the sensitivity analysis, and we recommend their inclusion. Multiple imputation is the best approach across all missingness mechanisms and offers some protection in some simple missing not at random contexts.
  13 in total

1.  Dealing with missing outcome data in randomized trials and observational studies.

Authors:  Rolf H H Groenwold; A Rogier T Donders; Kit C B Roes; Frank E Harrell; Karel G M Moons
Journal:  Am J Epidemiol       Date:  2011-12-23       Impact factor: 4.897

2.  Bias and efficiency of multiple imputation compared with complete-case analysis for missing covariate values.

Authors:  Ian R White; John B Carlin
Journal:  Stat Med       Date:  2010-12-10       Impact factor: 2.373

3.  Missing covariate data in medical research: to impute is better than to ignore.

Authors:  Kristel J M Janssen; A Rogier T Donders; Frank E Harrell; Yvonne Vergouwe; Qingxia Chen; Diederick E Grobbee; Karel G M Moons
Journal:  J Clin Epidemiol       Date:  2010-03-24       Impact factor: 6.437

4.  Bias and Precision of the "Multiple Imputation, Then Deletion" Method for Dealing With Missing Outcome Data.

Authors:  Thomas R Sullivan; Amy B Salter; Philip Ryan; Katherine J Lee
Journal:  Am J Epidemiol       Date:  2015-09-02       Impact factor: 4.897

5.  Using the outcome for imputation of missing predictor values was preferred.

Authors:  Karel G M Moons; Rogier A R T Donders; Theo Stijnen; Frank E Harrell
Journal:  J Clin Epidemiol       Date:  2006-06-19       Impact factor: 6.437

6.  Multiple imputation for missing data in epidemiological and clinical research: potential and pitfalls.

Authors:  Jonathan A C Sterne; Ian R White; John B Carlin; Michael Spratt; Patrick Royston; Michael G Kenward; Angela M Wood; James R Carpenter
Journal:  BMJ       Date:  2009-06-29

7.  Performance of statistical methods for meta-analysis when true study effects are non-normally distributed: A simulation study.

Authors:  Evangelos Kontopantelis; David Reeves
Journal:  Stat Methods Med Res       Date:  2010-12-09       Impact factor: 3.021

8.  Changes in Arterial Access Site and Association With Mortality in the United Kingdom: Observations From a National Percutaneous Coronary Intervention Database.

Authors:  Mamas A Mamas; James Nolan; Mark A de Belder; Azfar Zaman; Tim Kinnaird; Nick Curzen; Chun Shing Kwok; Iain Buchan; Peter Ludman; Evangelos Kontopantelis
Journal:  Circulation       Date:  2016-03-11       Impact factor: 29.690

9.  Eliciting and using expert opinions about dropout bias in randomized controlled trials.

Authors:  Ian R White; James Carpenter; Stephen Evans; Sara Schroter
Journal:  Clin Trials       Date:  2007       Impact factor: 2.486

Review 10.  Handling missing data in RCTs; a review of the top medical journals.

Authors:  Melanie L Bell; Mallorie Fiero; Nicholas J Horton; Chiu-Hsieh Hsu
Journal:  BMC Med Res Methodol       Date:  2014-11-19       Impact factor: 4.615

View more
  30 in total

1.  The Incredible Years Teacher Classroom Management Program: Outcomes from a Group Randomized Trial.

Authors:  Wendy M Reinke; Keith C Herman; Nianbo Dong
Journal:  Prev Sci       Date:  2018-11

2.  Associations between urinary biomarkers of oxidative stress in the third trimester of pregnancy and behavioral outcomes in the child at 4 years of age.

Authors:  Anna-Sophie Rommel; Ginger L Milne; Emily S Barrett; Nicole R Bush; Ruby Nguyen; Sheela Sathyanarayana; Shanna H Swan; Kelly K Ferguson
Journal:  Brain Behav Immun       Date:  2020-09-06       Impact factor: 7.217

3.  Prenatal exposure to airborne polycyclic aromatic hydrocarbons and childhood growth trajectories from age 5-14 years.

Authors:  Andrew G Rundle; Dympna Gallagher; Julie B Herbstman; Jeff Goldsmith; Darrell Holmes; Abeer Hassoun; Sharon Oberfield; Rachel L Miller; Howard Andrews; Elizabeth M Widen; Lori A Hoepner; Frederica Perera
Journal:  Environ Res       Date:  2019-07-19       Impact factor: 6.498

4.  Trajectories of Physical Work Capacity in Early Symptomatic Osteoarthritis of Hip and Knee: Results from the Cohort Hip and Cohort Knee (CHECK) Study.

Authors:  H J Bieleman; R Stewart; M F Reneman; W M van Ittersum; C P van der Schans; K W Drossaers-Bakker; F G J Oosterveld
Journal:  J Occup Rehabil       Date:  2019-09

5.  Multivariate Analysis of Metabolomic and Nutritional Profiles among Children with Autism Spectrum Disorder.

Authors:  Fatir Qureshi; James B Adams; Tapan Audhya; Juergen Hahn
Journal:  J Pers Med       Date:  2022-06-01

6.  Missing data in primary care research: importance, implications and approaches.

Authors:  Miguel Marino; Jennifer Lucas; Emile Latour; John D Heintzman
Journal:  Fam Pract       Date:  2021-03-29       Impact factor: 2.267

7.  Effectiveness of the baby-friendly community initiative on exclusive breastfeeding in Kenya.

Authors:  Elizabeth W Kimani-Murage; Judith Kimiywe; Antonina N Mutoro; Calistus Wilunda; Frederick Murunga Wekesah; Peter Muriuki; Bonaventure M Mwangi; Betty Mogesi Samburu; Nyovani Janet Madise; Stephen T McGarvey; Paula L Griffiths
Journal:  Matern Child Nutr       Date:  2021-02-02       Impact factor: 3.660

8.  Socioeconomic position over the life-course and subjective social status in relation to nutritional status and mental health among Guatemalan adults.

Authors:  Jithin Sam Varghese; Rachel Waford Hall; Ann M DiGirolamo; Reynaldo Martorell; Manuel Ramirez-Zea; Aryeh D Stein
Journal:  SSM Popul Health       Date:  2021-07-21

9.  Associations of Agency and Communion With Domain-Specific Self-Perceptions of Aging: A Cross-Sectional Study In Old-Old Adults in Poor Health.

Authors:  Anne Blawert; Sarah K Schäfer; Susanne Wurm
Journal:  Int J Aging Hum Dev       Date:  2021-10-18

10.  Incidence of uterine leiomyoma in relation to urinary concentrations of phthalate and phthalate alternative biomarkers: A prospective ultrasound study.

Authors:  Victoria Fruh; Birgit Claus Henn; Jennifer Weuve; Amelia K Wesselink; Olivia R Orta; Timothy Heeren; Russ Hauser; Antonia M Calafat; Paige L Williams; Donna D Baird; Lauren A Wise
Journal:  Environ Int       Date:  2020-12-21       Impact factor: 9.621

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.