Literature DB >> 30023456

Comparison of intent-to-treat analysis strategies for pre-post studies with loss to follow-up.

Wenna Xi1, Michael L Pennell1, Rebecca R Andridge1, Electra D Paskett2.   

Abstract

In pre-post studies when all outcomes are completely observed, previous studies have shown that analysis of covariance (ANCOVA) is more powerful than a change-score analysis in testing the treatment effect. However, there have been few studies comparing power under missing post-test values. This paper was motivated by the Behavior and Exercise for Physical Health Intervention (BePHIT) Study, a pre-post study designed to compare two interventions on postmenopausal women's walk time. The goal of this study was to compare the power of two methods which adhere to the intent-to-treat (ITT) principle when post-test data are missing: ANCOVA after multiple imputation (MI) and the mixed model applied to all-available data (AA). We also compared the two ITT analysis strategies to two methods which do not adhere to ITT principles: complete-case (CC) ANCOVA and the CC mixed model. Comparisons were made through analyses of the BePHIT data and simulation studies conducted under various sample sizes, missingness rates, and missingness scenarios. In the analysis of the BePHIT data, ANCOVA after MI had the smallest p-value for the test of the treatment effect of the four methods. Simulation results demonstrated that the AA mixed model was usually more powerful than ANCOVA after MI. The power of ANCOVA after MI dropped the fastest as the missingness rate increased; in most simulated scenarios, ANCOVA after MI had the smallest power when 50% of the post-test outcomes were missing.

Entities:  

Keywords:  ANCOVA; Missing data; Mixed model; Multiple imputation; Randomized trial

Year:  2018        PMID: 30023456      PMCID: PMC6022256          DOI: 10.1016/j.conctc.2018.05.008

Source DB:  PubMed          Journal:  Contemp Clin Trials Commun        ISSN: 2451-8654


Introduction

In a pre-post study a treatment is evaluated by measuring responses both before and after the study for each participant in a treatment group and a control group. Pre-post study designs have been widely used in clinical trials, psychology, education, and sociology. For example, our research was motivated by the Behavior and Exercise for Physical Health Intervention (BePHIT) Study, a pre-post study designed to compare two interventions intended to promote walking in postmenopausal women [1]. When there is complete follow-up, previous studies have shown that, in terms of testing the treatment effect, analysis of covariance (ANCOVA) is more powerful than a comparison of change scores [[2], [3], [4], [5]]. However, in reality, missing data, in particular loss to follow-up, is very common in pre-post studies. For instance, in the BePHIT study, 17% of the participants did not finish the study. With unbalanced sample sizes for pre- and post-test levels in each treatment group, a regular ANCOVA or change score analysis cannot be conducted without dropping any subjects. Therefore the most straightforward method for handling missing values is to exclude all the subjects with missing data. This type of analysis is called the complete-case (CC) analysis. The CC analysis is usually not recommended, since it throws away information collected in the study and does not follow the intent-to-treat (ITT) principle for clinical trials [6,7]. Nowadays, one popular way to handle missing data is multiple imputation (MI) [6]. For instance, in pre-post studies, missing follow-ups can be simulated multiple times using the baseline outcome value and measured covariates and the results of the analysis of each complete data set are combined to account for the uncertainty introduced by the imputations [8]. Another approach often used for data with repeated measures is the mixed model, where all available pre- and post-test values are regressed over treatment and timepoint indicators, assuming some variance-covariance structure for the repeated measures. The main goal of this study was to compare the power of two analysis methods which adhere to ITT principles: the mixed model and ANCOVA after MI for pre-post studies when missing post-test is present. We also wanted to compare these methods to two methods that do not adhere to an ITT principle: ANCOVA and the mixed model using only completely observed cases. These methods were first compared in the context of our motivating example (BePHIT) and then in simulation studies based on the BePHIT data. A parallel set of simulations comparing the type I error rate of the four methods were conducted as well.

Motivating example

The Behavior and Exercise for Physical Health Intervention (BePHIT) Study was a randomized controlled study of a 12-week walking intervention conducted on postmenopausal women between January 2008 and March 2009 [1]. The primary outcome was the change in time for women to finish a one-mile walk. In addition to one-mile walk time, anthropometric and psychometric measures were obtained at pre- and post-test. After passing the selection criteria, 71 participants were stratified by BMI and randomized into either a coach group or a no-coach group. For women in the coach group, a trained coach was assigned. The role of the coach was to explain the intervention, provide the first week's steps goal, train subjects to use a pedometer and the Interactive Voice Response (IVR) system to collect data, and offer help during the intervention. Women in the no-coach group received similar instructions, training, and help, except that they were not informed that they had access to a coach. Although both groups received a treatment, to be consistent with the terminology in this paper, we will refer to the coach condition as the treatment and the no coach condition as the control. Among the 71 randomized participants, 35 were assigned to the treatment group and 36 to the control group. For the control group, baseline walking time was only available for 35 patients. In total, 12 (17%) patients dropped out before the post walking test, 4 of whom were in the treatment group and 8 in the control group. The drop out rate did not differ significantly across the two groups (). The original study reported 2 withdrew and 12 did not complete in the treatment group, and 7 withdrew and 11 did not complete in the control group [1]. In that study, “completed” was defined as completing all post-test assessments within 30 days after the end of the walking intervention. These post-test assessments included the walk test and anthropometric and psychometric measures not considered in this paper. Those who had their post-test score recorded but did not finish all their anthropometric and psychometric measures were also included in this analysis, which led to fewer dropping out here.

Analysis methods for pre-post studies with complete data and missing data

Pre-post studies

A pre-post study is a randomized controlled study where outcome values are measured both before and after the study. As opposed to treatment-control studies where the outcome variable is only measured once, pre-post studies allow investigators to account for the level of the outcome variable before the treatment is applied. Different from a one-group pre-post design, a treatment-control pre-post study controls for secular trends [9,10]. In the BePHIT study, for instance, besides the intervention, the improvement of women's one-mile walk time may have been caused by some other factors, such as a national walking campaign, affecting the women during the same time period. A one-group pre-post study fails to consider these factors; however a treatment-control pre-post study accounts for secular trends by comparing the results from the treatment group to a control group observed over the same period of time. If we let be the randomization process, be the treatment process, and and be the pre- and post-test measure of a treated and control participant, respectively, then a pre-post study design can be illustrated by the following:

Analysis of pre-post studies with complete data

Many analysis approaches for pre-post studies have been discussed [2,3,5,9,11,12]. Arguably the two most common analysis methods are the change score analysis and Analysis of Covariance (ANCOVA) [2]. We discuss these two methods and their statistical power in the following.

Change score analysis

A change score analysis first obtains the difference in outcome values before and after the experiment, and then regresses the difference on the treatment assignment using the following model:where is the post-test outcome level for subject i, is the pre-test outcome level for subject i, is the indicator variable for treatment assignment, and is the error term for subject i (). Note that is the change score for subject i during the experiment. In the model above, quantifies the effect of treatment assignment on change in outcome level from pre to post. Since is a binary variable, the change score test is equivalent to a two-sample t-test comparing the mean of between treatment and control groups.

ANCOVA

Unlike the change score method, in ANCOVA, post-test value () is treated as the outcome variable and pre-test value () is treated as a predictor. The ANCOVA model can be expressed as:where and are as defined in the change-score model. ANCOVA assumes that pre-test values are measured without error [5]. This assumption holds for variables, such as weight and height, that can be measured precisely. However, it is often violated for self-reported measurements and educational or psychological tests. In the model above, is the effect of treatment assignment on the post-test scores adjusting for the pre-test scores. The ANCOVA model (2) can also be written aswhere from (2) [2,5]. Thus, ANCOVA can be viewed as an extension of the change score model (1) to include pre-test level as a predictor.

Power comparison: change score analysis vs. ANCOVA

Oakes and Feldman compared the detectable treatment effects of the change score and ANCOVA models [5]. Assume regardless of the experimental group, Number of subjects in each group is the same, and Pre-test is measured without error. Under the assumption of normally distributed errors, the detectable treatment effect for the change score analysis at type-I and type-II error rates of α and β isand the detectable treatment effect for ANCOVA iswhere , is the xth quantile of the standard normal distribution, and m is the number of subjects in each experimental group. Therefore we have Assuming , Thus, with complete data, the detectable treatment effect for ANCOVA is always less than or equal to the detectable treatment effect for the change score analysis. Also, from (4) we can see that the discrepancy in power increases as the correlation between pre- and post-test decreases.

Missing data

Both the change score analysis and ANCOVA require complete follow-up of subjects. Unfortunately, missing post-test data are very common in pre-post studies since participants may drop out because they moved, are unsatisfied with their performance in the study, etc. For instance, in the BePHIT study, 17% of the patients dropped out prior to study completion. Post-test data could be missing completely at random, at random, or not at random [6]. Missing completely at random (MCAR) means the missing data does not depend on anything in the data collecting process (e.g., data missing due to a data entry error). We say a study has data missing at random (MAR) when the missingness only depends on observed data, but not those data that are missing (or would have been observed). A subject being removed from a study due to an unhealthy pre-test is an example of MAR. When MAR does not hold, we say that data are not missing at random (NMAR). For example, data would be NMAR if subjects drop out because they know their post-test outcome will be poor.

Statistical methods for missing data

Commonly used methods for handling missing data include: complete-case (CC) analysis, weighting adjustments, imputation, and model-based methods [6,13,14]. We focus in this paper on three natural approaches to handling loss to follow-up in a pre-post study. First, one can remove the data for subjects with post-test missing and carry out a complete-case (CC) analysis. Second, the change score model can be generalized to a mixed model, in which unequal sample sizes for pre- and post-test levels is allowed. Finally, ANCOVA could be applied following imputation of the missing data. These three approaches are summarized below.

Complete-case (CC) analysis

The most straightforward way to deal with missing data is to delete all subjects with any missing observations, no matter if data were partially collected or not. The dataset obtained after deleting all subjects with missing values is called a complete-case dataset and the analysis of these data is called a complete-case (CC) analysis. The CC analysis is simple to conduct and unbiased under MCAR. In certain situations, it is also unbiased under MAR. For instance, when estimating the regression model of Y on where Y is incompletely observed, the estimation conditions on the values of the X's. Thus, the CC analysis is unbiased if the missingness only depends on the X's but not the Y [6,14]. As a result, the CC ANCOVA model is unbiased if the missingness of post-test values only depends on the pre-test values. In addition to the possibility of bias, another disadvantage of the CC analysis is that information is thrown away by deleting subjects. Estimation based on a CC analysis may result in larger variances than methods for incomplete data. Also, since CC analysis does not include all subjects randomized in the final analyses, it does not adhere to the intent-to-treat (ITT) principle of clinical trials [7]. It is for these reasons that CC analyses are usually not employed in intervention studies.

Mixed models

The mixed model is a regression model that includes both population-level and subject-level effects. It assumes responses are MAR. For a pre-post study with one treatment group and one control group, a mixed model can be written as: Here, is the response of subject i at time point j (), and are indicators of treatment group and post-test, respectively, and is a random subject effect (), which is independent of the random error (). Note that if we subtract the pre- and post-test outcomes we getwhere . Thus, when there is no missing data, the maximum likelihood estimate of is equivalent to the treatment effect () in the change score model (1). However, unlike the change score model, the mixed model can include data on subjects with just one of the two outcome values. An alternative formulation of the mixed model may also be used to analyze pre-post data:where , , , and for . The major practical difference between (5) and (6) is that (5) assumes a positive correlation between pre- and post-measures (which is expected in a pre-post study) while (6) permits positive and negative correlations. Expression (6) is often referred to as a marginal formulation of a mixed model since the regression function does not include subject-specific effects. Sullivan et al. consider a different mixed model for studies with missing follow-up in which post-test and an auxiliary outcome (potentially the main outcome measured at an earlier time point following randomization) are modeled using a linear mixed model containing fixed effects of treatment and a baseline covariate (potentially the pre-test value of the outcome). Though a useful approach in some settings, the Sullivan et al. model cannot be applied in studies with a single post-test and loss-to-follow-up and hence will not be considered further.

Multiple imputation (MI)

The general idea of multiple imputation is to fill in the missing data several times and obtain several “complete” data sets, then conduct the statistical inferences based on those completed data sets [6,15]. Compared to single imputation, multiple imputation accounts for variability in the estimate for the missing value. The general procedure for MI can be summarized as three steps: Step 1: Fill in the missing data N times and get N complete data sets. Step 2: Analyze each of the N data sets separately using analysis methods for complete data. Step 3: Use Rubin's rules to combine the analysis results for the N data sets and obtain the final result [16]. For pre-post data with only post-test values missing, a regression model of on pre-test () and some selected covariates is first estimated using all observations with observed values of [17]: At imputation k (), new coefficients are sampled from the posterior predictive distribution and then the missing value is sampled from , where are the values of the covariates of subject i in the imputation model. Consider a linear regression modelwhere is a length n vector of outcomes, is an covariate matrix, is a length p vector of regression parameters, and is a length n vector of errors (). Assume the outcomes are partially observed, the covariates are completely observed, the missing data mechanism in is MAR, and that is known. Suppose we are interested in estimating . Let and be the complete-case (CC) ML estimator and MI estimator of , respectively. It has been shown that [18]: Thus, both the CC ML and MI estimates are unbiased, but their variances differ. Since is positive-definite, we see that the variance of the MI estimator is always larger than the variance of the ML estimator using only complete cases. The above results can be generalized to unknown [18]. Thus, when analyzing pre-post data using the ANCOVA model, estimates from both CC and MI analyses are unbiased, but the standard errors from the CC analysis should be smaller.

Analysis of the BePHIT data

Analyses

Logistic regression was used to determine whether or not dropping out was related to any pre-test measures. The estimated odds ratios (ORs) from univariate logistic models and their p-values are listed in Table 1. Only waist-hip ratio was associated with missing post-test (p = 0.04). People with larger waist-hip ratio were more likely to drop out of the study. Since waist-hip ratio was related to the dropout rate, the missing data mechanism was not MCAR.
Table 1

Logistic regression analysis of drop out in BePHIT.

VariableCoefficientSEOdds Ratiop-value
Design and Outcome
 Treatment (Coach)−0.830.670.440.21
 Pre-test walk time0.110.161.120.47
 Baseline Anthropometrics
 Pulse rate0.040.031.050.12
 Waist/hip (+0.1 units)1.230.593.420.04
 BMI0.110.081.120.16
Baseline Psychometrics†
 Negative exercise thoughts0.320.521.380.54
 Exercise stage of change0.210.421.230.62
 Social support from family−0.160.480.860.75
 Social support from friends−0.940.710.390.19
 Walking Self-efficacy−0.170.170.840.32
 Self-efficacy to walk 30 min0.190.211.210.38
 Exercise goals0.120.361.120.74
 Exercise planning−0.360.540.700.50

† Measures under this category are all quantitive scores. More details about these measures can be found in Ref. [1].

Logistic regression analysis of drop out in BePHIT. † Measures under this category are all quantitive scores. More details about these measures can be found in Ref. [1]. David et al. used linear mixed models to analyze the pre-post changes [1]. Here we analyzed the data using a mixed model (with all available data), ANCOVA after MI, and the corresponding CC analyses (the mixed model with CC and ANCOVA with CC). Results from each method were compared. The ANCOVA model is valid here since the primary outcome, time for one-mile walk, can be assumed to be measured without error. For both the all-available-data mixed model and ANCOVA after MI, all 70 participants with a baseline walking time were included. One subject in the control group was excluded due to missing both pre- and post-test scores. For the two CC analyses, only participants with both baseline and post-intervention walking time were used (). We used the marginal formulation of the mixed model (6) and the Kenward-Roger method was used for computing denominator degrees of freedom [[19], [20], [21]]. In MI, post-test values were imputed using a model that included the 12 baseline measures listed in Table 1, including treatment. To preserve the reproducibility of the imputation, missing values were imputed 20 times based on a rule of thumb proposed by White et al. [15]. Imputed data were analyzed using the ANCOVA model (2) and results were combined using the standard MI combining rules [6,15]. The imputation model was fit using the completely observed cases to check whether the variables were predictive of post-test. The estimates from the imputation model are listed in Table 2. The of the imputation model was 0.6975. However, only two variables, pre-test walk time and BMI, were significant at the 0.05 level.
Table 2

Regression Estimates for Imputation Model in BePHIT analysis.

VariableEstimateSEp-value
 Intercept−3.074.230.47
Design and Outcome
 Treatment (Coach)0.010.400.98
 Pre-test walk time0.780.11<0.0001
Baseline Anthropometrics
 Pulse rate−0.010.020.61
 Waist/hip2.564.010.52
 BMI0.160.05<0.01
Baseline Psychometrics
 Negative exercise thoughts−0.280.390.48
 Exercise stage of change0.150.280.60
 Social support from family0.290.300.35
 Social support from friends−0.260.360.48
 Walking Self-efficacy−0.130.140.37
 Self-efficacy to walk 30 min0.140.140.31
 Exercise goals0.290.340.41
 Exercise planning−0.540.550.33
Regression Estimates for Imputation Model in BePHIT analysis.

Results

Estimates and tests of a treatment effect for the four methods are summarized in Table 3. All four methods showed no coach effect on the change in one-mile walk time. This result is consistent with those reported by David et al. [1]. As we can see, the results from the two mixed models were very similar. However, there were obvious differences in both the estimates and hypothesis tests between the two ANCOVA analyses. When MI was used, the estimated effect for treatment changed from −0.08 to −0.24, and its corresponding standard error changed from 0.38 to 0.63.
Table 3

Comparison of the BePHIT treatment effects from the four analysis methods.

MethodVariableEstimateSEtp-value
Mixed Model, CCTrt. × Post−0.0480.38−0.130.8995
Mixed Model, AATrt. × Post−0.0490.38−0.130.8959
ANCOVA, CCTreatment−0.0840.38−0.220.8251
ANCOVA, MITreatment−0.2420.63−0.390.7005

Abbreviations: CC, complete-case analysis; AA, all-available-case analysis.

Comparison of the BePHIT treatment effects from the four analysis methods. Abbreviations: CC, complete-case analysis; AA, all-available-case analysis. ANCOVA after MI gave the most significant test of treatment among all four methods. The treatment effect tests from the two ANCOVA methods were more significant than from the two mixed models. Within each model, the test of treatment effect using all available cases was more significant than using complete cases only.

Simulation study

Data generation

Simulation studies were conducted to compare the power and the type I error rate of the ITT and complete-case analysis methods. The coefficient values used in the data generation process were, for the most part, based on the results from the BePHIT analysis. The data sets used to compare power were generated from the following model:; ; ; . Here is the outcome of subject i at time j and and are indicator variables of post-test and treatment, respectively. The variable is the waist-hip ratio of subject i, which was generated from to mimic the distribution in the BePHIT data. Under our data generation model, which was the value in the BePHIT data. Although the estimated main effect of treatment was not 0 in the BePHIT analysis, it was set to zero in (8) so that the expected outcome value of the two groups was the same at baseline. Also, a main effect of waist-hip ratio was added because waist-hip ratio was significantly related to missing follow-ups in the BePHIT study. Its coefficient was adjusted to −2 so that waist-hip ratio was, on average, significantly associated with the outcome. In the original analyses of the BePHIT study, the interaction effect of treatment and time was not significant. However, in order to conduct power comparisons in the simulation studies, the coefficient for this interaction effect was adjusted to −1 so that, on average, it was significant for all the analysis models. The model used to generate the data for the type I error rate comparisons was similar to (8), except that the interaction between treatment and time was set to 0. Two different group sizes were considered: (i.e., the BePHIT sample size) and per group (). For each group size, 500 data sets were simulated. Post data were then set to be missing at rates of 20%, 30%, 40%, and 50% missingness. At each percentage of missingness, missing data were generated as missing completely at random (MCAR) and missing at random (MAR). For MAR, missingness was generated under four different conditions: dependent on waist-hip ratio, dependent on both waist-hip ratio and pre-test level, dependent on waist-hip ratio, pre-test level, and treatment, and dependent on waist-hip ratio, pre-test level, treatment, and the interaction between pre-test level and treatment. To generate the missing data, a Bernoulli indicator was drawn for each subject with the probability defined by the following logistic regression model: The γ′s were first estimated using the BePHIT data. However, since the analyses of the drop out rate in the original study showed very different effects of the three main effect terms (Table 1), the coefficients of and were adjusted so that their effects would be similar to the pre-test level's. The coefficient of was adjusted by forcing the odds ratio of drop out to be two for a one standard deviation change in waist-hip ratio. The coefficient of was adjusted by forcing the odds ratio of drop out for the treatment group versus the control group to be four. In the last simulated scenario, MAR dependent on waist-hip ratio, pre-test level, treatment, and the interaction between pre-test level and treatment, the coefficient of was adjusted such that the average effect of between the two groups was the same as in the previous scenarios, and the coefficient of was adjusted such that the average treatment effect was the same as in the previous scenarios. Our choice of main effects and interaction coefficient resulted in control subjects with smaller being more likely to drop out and intervention subjects with larger being more likely to drop out. The γ′s were set to 0 when the missingness did not depend on their corresponding variables. Finally, was adjusted for each percentage of missingness. Table 4 provides the values of the coefficients for each scenario.
Table 4

Coefficients for log-odds of missing follow-up used in the simulation study.

% MissingnessMissing Mechanismγ0γ1γ2γ3γ4
20MCAR−1.3860000
MAR (WH)−3.2521.989000
MAR (WH + Pre)−4.9101.9890.09400
MAR (WH + Pre + Trt)−5.6031.9890.0941.3860
MAR (WH + Pre + Trt + Pre × Trt)−3.2511.989−0.039−3.3190.265
30MCAR−0.8470000
MAR (WH)−2.7131.989000
MAR (WH + Pre)−4.3711.9890.09400
MAR (WH + Pre + Trt)−5.0641.9890.0941.3860
MAR (WH + Pre + Trt + Pre × Trt)−2.7121.989−0.039−3.3190.265
40MCAR−0.4050000
MAR (WH)−2.2711.989000
MAR (WH + Pre)−3.9291.9890.09400
MAR (WH + Pre + Trt)−4.6221.9890.0941.3860
MAR (WH + Pre + Trt + Pre × Trt)−2.2701.989−0.039−3.3190.265
50MCAR00000
MAR (WH)−1.8661.989000
MAR (WH + Pre)−3.5241.9890.09400
MAR (WH + Pre + Trt)−4.2171.9890.0941.3860
MAR (WH + Pre + Trt + Pre × Trt)−1.8641.989−0.039−3.3190.265

Abbreviations: MAR (WH), missingness dependent on waist-hip ratio; MAR (WH + Pre), missingness dependent on waist-hip ratio and pre-test level; MAR (WH + Pre + Trt), missingness dependent on waist-hip ratio, pre-test level, and treatment assignment; MAR (WH + Pre + Trt + Pre Trt), missingness dependent on waist-hip ratio, pre-test level, treatment, and the interaction between pre-test level and treatment.

Coefficients for log-odds of missing follow-up used in the simulation study. Abbreviations: MAR (WH), missingness dependent on waist-hip ratio; MAR (WH + Pre), missingness dependent on waist-hip ratio and pre-test level; MAR (WH + Pre + Trt), missingness dependent on waist-hip ratio, pre-test level, and treatment assignment; MAR (WH + Pre + Trt + Pre Trt), missingness dependent on waist-hip ratio, pre-test level, treatment, and the interaction between pre-test level and treatment. For each scenario mentioned above, simulated data sets were analyzed by both mixed models and ANCOVA after multiple imputation (MI). Since outcome values were generated from a model containing waist-hip ratio, in addition to the mixed model and ANCOVA model mentioned in Section 3, each method was also conducted with waist-hip ratio in the analysis model. For comparison purposes, we also performed complete-case (CC) analyses using the same mixed models and ANCOVA models. The mixed model used for the analysis was:where is as defined in (6). Similar to the BePHIT study analysis, the Kenward-Roger method was used for computing denominator degrees of freedom [[19], [20], [21]]. The ANCOVA model used for the analysis was:where . When waist-hip ratio was not included in the analysis model, in the mixed model (10) and in the ANCOVA model (11) were set to 0. For MI, missing follow-ups were imputed 20 times using the following imputation model: That is, the predictors in the imputation model were all the main effects that were used in generating the data and all their two-way interactions. The statistical software used for both data generation and analyses was SAS 9.4 (SAS Inc., Cary, NC). The DATA step was used to generate the data, the REG procedure was used to fit the ANCOVA model, the MI and MIANALYZE procedures were used to do the multiple imputation, and the MIXED procedure was used to fit the mixed model. Fig. 1 compares the power of the tests of a treatment effect for each method when waist-hip ratio was included in the analysis models and subjects per group. Here α was set to 0.05. The power of each method decreased as the percentage of missingness increased. The complete-case (CC) ANCOVA was the best in terms of the power when percentage of missingness was smaller (20% and 30%). When missing proportion became larger (40% and 50%), the all-available-case mixed model yielded similar results to CC ANCOVA. Under MCAR, ANCOVA after multiple imputation (MI) was slightly more powerful than the mixed model using all available data when the percentage of missingness was smaller (20% and 30%). Under MAR, ANCOVA after MI and the mixed model using all available data were still comparable when post data were missing at 20%. However, the power of ANCOVA after MI decreased dramatically as the percentage of missingness increased and for most scenarios was the least powerful method when 50% of the post data were missing. When the true logit model for drop out included an interaction between treatment and pre-test (), the power of the CC mixed model was considerably lower than all other methods when 20–40% were missing follow-up; when 50% were missing follow-up, the power of the CC mixed model was comparable to the power of ANCOVA after MI. When per group, the power of each test was very high (mostly ) and CC ANCOVA and the all-available-case mixed model had very similar power under each scenario (Fig. S1 in the Supplementary Materials). Similar trends in power were observed when waist-hip ratio was omitted from the analysis model (Figs. S2 and S3 in the Supplementary Materials).
Fig. 1

Power comparisons for analysis models including WH and per group.

Power comparisons for analysis models including WH and per group. Coefficient estimates and standard errors of the power simulations are provided in Table S1 – S8 in the Supplementary Materials. For the most part, all four methods produced unbiased estimates. The one exception to this trend was when the CC mixed model was applied to data with missingness dependent on an interaction between treatment and pre-test; in that scenario the treatment-time interaction effect (i.e., the treatment effect of interest) was biased (between 6 and 13% across missingess rates). The all-available-case mixed model consistently had smaller standard errors of the estimates than the CC mixed model, while the reverse was true for the ANCOVA analyses due to the relationship defined in (7). Also, the estimates were more accurate and precise at the larger sample size ( per group). Fig. 2 compares the type I error rate of the tests of a treatment effect for each method when per group and waist-hip ratio was included in the analysis models. Type-I error rates were similar under MCAR and when missingness only depended on WH, though in the latter scenario the type-I error rate for CC ANCOVA was inflated under 50% missingness. When missingness depended on treatment alone or treatment and pre-test, the mixed models generally exhibited deflated type-I error rates and with the exception of one case (ANCOVA after MI when missingess depended on WH, pre-test, treatment, and the pre-test by treatment interaction) the type-I error rates for the ANCOVA models were close to the nominal level. When there was interaction between treatment and pre-test in the missingness model, type-I error rates were deflated under ANCOVA after MI at the higher missingness rates and were inflated under the mixed models when 50% were missing post-test. Differences in type-I error rate were more variable when per group with no single method consistently providing the lowest or highest type-I error rates (Fig. S4 in the Supplementary Materials). Similar trends in type-I error rate were observed when WH was excluded from the regression models (Figs. S5 and S6 in the Supplementary Materials).
Fig. 2

Type I Error Rate Comparisons for Analysis Models with WH and n = 35.

Type I Error Rate Comparisons for Analysis Models with WH and n = 35.

Discussion

Summary

In the BePHIT data analyses, p-values were smaller for ANCOVA than for mixed models when testing the treatment effects. Also, for both ANCOVA and mixed models, p-values were smaller for ITT analyses (ANCOVA after MI and the AA mixed model) than CC analyses. However, the treatment effects estimated by ANCOVA after MI and CC ANCOVA were remarkably different (−0.08 for CC vs. −0.24 for MI), as were the standard errors (0.38 for CC vs. 0.63 for MI). It is also interesting that the p-value from CC ANCOVA was smaller than the p-value from the all-available-data mixed model. Thus, for the BePHIT data, the benefit of switching the CC mixed model to the CC ANCOVA model outweighed the benefit of switching the CC mixed model to the all-available-data mixed model. The simulation studies showed that all four methods, ANCOVA after multiple imputation (MI), ANCOVA for complete cases (CC), the all-available-case mixed model, and the CC mixed model, generally produced unbiased estimates of the treatment effect when waist-hip ratio was included in the analysis model. The only exception was the CC mixed model under MAR dependent on waist-hip ratio, pre-test value, treatment, and the interaction between pre-test value and treatment. In this scenario, the treatment effect was confounded by pre-test in complete case analyses since the presence of an interaction in the missingness model resulted in an inverse association between assignment to treatment and pre-test. When there is no missing data, the mixed model (5) is equivalent to the change score model (1), which does not account for pre-test. Thus, it is not surprising that the complete case change score analysis produced a biased treatment effect while the CC ANCOVA analysis provided an unbiased treatment effect since it adjusts for pre-test. This result implies that among CC analyses, treatment effect estimates from an ANCOVA model are more robust to missingness pattern than the change score analysis. As expected, the simulation studies showed that ANCOVA was more powerful than the mixed model for CC analyses, and the mixed model was generally more powerful when all available data were used. Also, since CC ANCOVA provides more precise estimates than ANCOVA after MI [18] and ANCOVA is more powerful than the mixed model [5], it was not surprising that CC ANCOVA had the largest power under almost all scenarios. In some situations, the difference in power between the ANCOVA and mixed model analyses could be explained, in part, by the mixed models exhibiting type-I error rates below nominal levels, but this was not true in general. Also, the benefit in power from CC ANCOVA generally did not come at the cost of an inflated type-I error rate. The results from the simulation studies were not consistent with the analyses of the BePHIT data. The simulation studies suggested that ANCOVA after MI was almost always the least powerful method for testing the treatment effect, while this approach resulted in the smallest p-value in the BePHIT analysis. This discrepancy may be the result of using different imputation models between the simulation studies and the BePHIT analyses. The imputation model in the simulations included only the variables used in generating the data and their two-way interactions, while the imputation model for the BePHIT studies also included some baseline anthropometric and psychometric measures. This resulted in a difference in the of the imputation model for the two studies: for the BePHIT study, while the average was smaller for each simulated scenario when 20% of the data were missing (between 0.61 and 0.65).

Comparison with similar studies

Although many papers have compared analysis methods for pre-post studies when data are completely observed [[2], [3], [4], [5],22] considerably fewer have compared methods under missing post-test values. Liu et al. [23] and Hyer and Waller [24] compared mixed models to ANCOVA models applied to complete cases. Mehrotra et al. imputed missing post-test values, applied a robust regression model to the pre-post changes, and compared the approach to more standard analyses for missing longitudinal data (weighted generalized estimating equations and mixed models) [25]. Mehrotra et al.’s work was similar to ours in that they compared ITT analysis methods, one of which involved imputation. However, the authors did not consider ANCOVA models which many consider the preferred analysis approach for complete data [5]. Furthermore, Mehotra et al. considered studies with several follow-ups and thus imputations were based on more observable outcome data than in two group pre-post studies. Sullivan et al. considered missing data in a univariate outcome, a multivariate outcome, and a baseline covariate and compared MI, applied both overall and by randomized group, with a complete case analysis and a mixed model. While there are some similarities between our study and Sullivan et al.'s, there are two key differences. First, Sullivan et al. only considered a large sample size in their simulations ( total). While large sample sizes such as these may be common in clinical trials, many studies with pre-post designs involve much fewer subjects. For instance, our motivating study BePHIT consisted of 35 women per treatment group. Another example is a recent crossover study on depression and healthier food choices, where a total of 58 women were recruited [26]. Thus, motivated by the BePHIT study, we considered scenarios with small sample sizes. The second major difference between our study and Sullivan et al.’s is the type of mixed model considered in the ITT analysis; we considered a model for pre-test and post-test with a time-dependent treatment effect while Sullivan et al. considered a mixed model for post-test and an auxiliary outcome adjusting for pre-test. Our mixed model provides a treatment effect on the pre-post change in the outcome and thus is an extension of the commonly used change-score analysis to accommodate missing data. From our experience, the Sullivan et al. model is not as common since, in a study with a single post-randomization follow-up, if a subject is lost-to-follow-up, they will not have any information on the main outcome or the auxiliary outcome. Despite the differences in the settings of the simulations, Sullivan et al. also observed that complete case analysis (CCA) was more efficient than multiple imputation (MI) for pre-post studies [27]. When there were multiple follow-ups, Sullivan et al. found that the mixed model was more efficient than MI, though MI outperformed CCA since the intermediate outcome was ignored in CCA but used in the imputation model.

Limitations

The simulation settings were based on the BePHIT data and thus may not be reflective of studies whose outcome variables and covariates come from different distributions or with different designs (e.g., unequal numbers in each treatment group). Only four missing data mechanisms were considered in the simulations. These scenarios were chosen based on the analyses of the BePHIT study. However, many other missing mechanisms are possible in biomedical studies. For example, if the subject dropped out the study because she knew that her post-test one-mile walk would be very bad, the data would be not missing at random (NMAR). However, this scenario, along with many other scenarios of missing data were not considered.

Recommendations

Even though CC ANCOVA appeared to be the most powerful method, it is usually not recommended since it does not adhere to the ITT principle of clinical trials [7]. In most simulated scenarios, the all-available-data (AA) mixed model provided a power benefit over ANCOVA after MI. Even in those scenarios where the power of ANCOVA after MI was larger than the power of the AA mixed model, the two powers were very close (e.g., Fig. 1(a) at 20 and 30% missingness). Besides the power loss in ANCOVA after MI, ANCOVA requires no measurement error in the outcome [5] and MI is more time consuming than fitting a mixed model. Therefore, for ITT analyses of pre-post studies with loss to follow-up, the AA mixed model is recommended.
  13 in total

1.  Statistical power for nonequivalent pretest-posttest designs. The impact of change-score versus ANCOVA models.

Authors:  J M Oakes; H A Feldman
Journal:  Eval Rev       Date:  2001-02

2.  Bias and efficiency of multiple imputation compared with complete-case analysis for missing covariate values.

Authors:  Ian R White; John B Carlin
Journal:  Stat Med       Date:  2010-12-10       Impact factor: 2.373

3.  Missing data in clinical studies: issues and methods.

Authors:  Joseph G Ibrahim; Haitao Chu; Ming-Hui Chen
Journal:  J Clin Oncol       Date:  2012-05-29       Impact factor: 44.544

4.  Comparison of data analysis strategies for intent-to-treat analysis in pre-test-post-test designs with substantial dropout rates.

Authors:  Agus Salim; Andrew Mackinnon; Helen Christensen; Kathleen Griffiths
Journal:  Psychiatry Res       Date:  2008-08-20       Impact factor: 3.222

5.  Should baseline be a covariate or dependent variable in analyses of change from baseline in clinical trials?

Authors:  Guanghan F Liu; Kaifeng Lu; Robin Mogg; Madhuja Mallick; Devan V Mehrotra
Journal:  Stat Med       Date:  2009-09-10       Impact factor: 2.373

6.  Small sample inference for fixed effects from restricted maximum likelihood.

Authors:  M G Kenward; J H Roger
Journal:  Biometrics       Date:  1997-09       Impact factor: 2.571

7.  Disentangling secular trends and policy impacts in health studies: use of interrupted time series analysis.

Authors:  Utz J Pape; Christopher Millett; John Tayu Lee; John T Lee; Josip Car; Azeem Majeed
Journal:  J R Soc Med       Date:  2013-04       Impact factor: 5.344

8.  A walking intervention for postmenopausal women using mobile phones and Interactive Voice Response.

Authors:  Prabu David; Janet Buckworth; Michael L Pennell; Mira L Katz; Cecilia R DeGraffinreid; Electra D Paskett
Journal:  J Telemed Telecare       Date:  2011-11-03       Impact factor: 6.184

9.  Multiple imputation using chained equations: Issues and guidance for practice.

Authors:  Ian R White; Patrick Royston; Angela M Wood
Journal:  Stat Med       Date:  2010-11-30       Impact factor: 2.373

10.  Depression, daily stressors and inflammatory responses to high-fat meals: when stress overrides healthier food choices.

Authors:  J K Kiecolt-Glaser; C P Fagundes; R Andridge; J Peng; W B Malarkey; D Habash; M A Belury
Journal:  Mol Psychiatry       Date:  2016-09-20       Impact factor: 15.992

View more
  12 in total

1.  A comparison of two different refeeding protocols and its effect on hand grip strength and refeeding syndrome: a randomized controlled clinical trial.

Authors:  Sissel Urke Olsen; Karin Hesseberg; Anne-Marie Aas; Are Hugo Pripp; Anette Hylen Ranhoff; Asta Bye
Journal:  Eur Geriatr Med       Date:  2021-06-04       Impact factor: 1.710

2.  Modest Sodium Reduction Increases Circulating Short-Chain Fatty Acids in Untreated Hypertensives: A Randomized, Double-Blind, Placebo-Controlled Trial.

Authors:  Li Chen; Feng J He; Yanbin Dong; Ying Huang; Changqiong Wang; Gregory A Harshfield; Haidong Zhu
Journal:  Hypertension       Date:  2020-06-01       Impact factor: 10.190

3.  A randomized control trial of Expand Your Horizon: An intervention for women with weight bias internalization.

Authors:  Alexandria E Davies; C Blair Burnette; Scott G Ravyts; Suzanne E Mazzeo
Journal:  Body Image       Date:  2021-12-24

4.  Sodium Reduction, Metabolomic Profiling, and Cardiovascular Disease Risk in Untreated Black Hypertensives.

Authors:  Li Chen; Feng J He; Yanbin Dong; Ying Huang; Gregory A Harshfield; Haidong Zhu
Journal:  Hypertension       Date:  2019-05-13       Impact factor: 9.897

5.  Physical fitness in older women with osteoporosis and vertebral fracture after a resistance and balance exercise programme: 3-month post-intervention follow-up of a randomised controlled trial.

Authors:  Brita Stanghelle; Hege Bentzen; Lora Giangregorio; Are Hugo Pripp; Dawn A Skelton; Astrid Bergland
Journal:  BMC Musculoskelet Disord       Date:  2020-07-18       Impact factor: 2.362

6.  A two-component intervention to improve hand hygiene practices and promote alcohol-based hand rub use among people who inject drugs: a mixed-methods evaluation.

Authors:  Salim Mezaache; Laélia Briand-Madrid; Linda Rahni; Julien Poireau; Fiona Branchu; Khafil Moudachirou; Yourine Wendzinski; Patrizia Carrieri; Perrine Roux
Journal:  BMC Infect Dis       Date:  2021-02-25       Impact factor: 3.090

7.  A Window Into Mental Health: Developing and Pilot-Testing a Mental Health Promotion Intervention for Mexican Immigrants Through the Ventanilla de Salud Program.

Authors:  Delia Lilian Martínez Rodríguez; Tonatiuh González Vázquez; Margarita Márquez Serrano; Mary de Groot; Alicia Fernandez; Ines Gonzalez Casanova
Journal:  Front Public Health       Date:  2022-04-15

8.  The effects of magnesium supplementation on abnormal uterine bleeding, alopecia, quality of life, and acne in women with polycystic ovary syndrome: a randomized clinical trial.

Authors:  Mahsima Jaripur; Hatav Ghasemi-Tehrani; Gholamreza Askari; Mahsa Gholizadeh-Moghaddam; Cain C T Clark; Mohammad Hossein Rouhani
Journal:  Reprod Biol Endocrinol       Date:  2022-08-02       Impact factor: 4.982

9.  Dietary Approaches to Stop Hypertension Dietary Intervention Improves Blood Pressure and Vascular Health in Youth With Elevated Blood Pressure.

Authors:  Sarah C Couch; Brian E Saelens; Philip R Khoury; Katherine B Dart; Kelli Hinn; Mark M Mitsnefes; Stephen R Daniels; Elaine M Urbina
Journal:  Hypertension       Date:  2020-11-16       Impact factor: 10.190

10.  Effectiveness of a web-based computer-tailored intervention promoting physical activity for adults from Quebec City: a randomized controlled trial.

Authors:  François Boudreau; Gilles R Dagenais; Hein de Vries; Michel Jean Louis Walthouwer; José Côté; Ginette Turbide; Anne-Sophie Bourlaud; Paul Poirier
Journal:  Health Psychol Behav Med       Date:  2020-12-08
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.