Literature DB >> 20963770

One-stage parametric meta-analysis of time-to-event outcomes.

F Siannis1, J K Barrett, V T Farewell, J F Tierney.   

Abstract

Methodology for the meta-analysis of individual patient data with survival end-points is proposed. Motivated by questions about the reliance on hazard ratios as summary measures of treatment effects, a parametric approach is considered and percentile ratios are introduced as an alternative to hazard ratios. The generalized log-gamma model, which includes many common time-to-event distributions as special cases, is discussed in detail. Likelihood inference for percentile ratios is outlined. The proposed methodology is used for a meta-analysis of glioma data that was one of the studies which motivated this work. A simulation study exploring the validity of the proposed methodology is available electronically.
Copyright © 2010 John Wiley & Sons, Ltd.

Entities:  

Mesh:

Year:  2010        PMID: 20963770      PMCID: PMC3020327          DOI: 10.1002/sim.4086

Source DB:  PubMed          Journal:  Stat Med        ISSN: 0277-6715            Impact factor:   2.373


1. Introduction

Meta-analysis of data from multiple studies of the same research question has achieved a very high profile in medical research over the recent years. Currently, particular attention is being given to the potential value of individual patient data (IPD) 1 and the need to handle the challenges of meta-analysis of time-to-event outcomes. Aggregate or summary data, such as hazard ratios and confidence intervals, can be used for time-to-event outcomes and they are commonly available in published papers. Methods for synthesizing evidence of this type (see discussion in 2), are borrowed from the methods used for summary statistics for simpler outcomes. However, in addition to the need for caution when extracting summary statistics of interest from papers or reports where they may not be clearly presented 3, this leaves little opportunity to examine the many characteristics of time-to-event data that may influence the results of standard analyses. To deal with this restriction, Fiocco et al. 4 have reconstructed data from the literature and provided a way to examine time-varying hazard models, an important generalization of what is normally possible with summary data. Other aspects of time-to-event data such as covariate adjustment may however be less easily handled with this approach. Thus, while IPD is considered the gold standard in meta-analysis in general 1, as all the relevant data are utilized, and approximations needed for aggregate data meta-analyses are avoided, their use is even more to be preferred with time-to-event outcomes for which a variety of distributional aspects may be of interest. Simmonds et al. review methods used in the meta-analysis of IPD from randomized trials 5 and Tudur-Smith et al. explore the heterogeneity of IPD meta-analysis using hierarchical Cox regression models 6. The logarithm of the hazard ratio (logHR) is the most prevalent summary measure used in the meta-analysis of time-to-event endpoints. Although some argue that it is always justified to consider the logHR with time-to-event data, this approach is most natural in the presence of a proportional hazards (PH) structure 7. However, in a meta-analysis, the PH assumption can be particularly restrictive, since it is imposed on multiple studies. Fiocco et al. have provided a means to consider time-varying hazard ratios but there remains scope to consider the potential value of other approaches. Here, the use of parametric models for meta-analysis of time-to-event IPD is explored as an alternative to the widely used Cox's PH model. Greater flexibility in the representation of treatment effects may be one advantage. Depending on the choice of model, various data structures can be naturally incorporated, with the accelerated failure time (AFT) structure being the most obvious alternative to the PH one. In principle, the combination of quite different data structures is possible since likelihoods of different forms from multiple studies can be combined to provide a basis for inference8. In addition, the use of a parametric model allows straightforward incorporation of covariates. If we do not want to only consider models with a PH structure, the logHR cannot be adopted as the target of inference. As an alternative, we propose the use of a convenient ratio of percentiles, typically related to two treatment groups being compared, and which has the added advantage that it is defined for all distributions. An obvious choice is the median ratio. More generally, the percentile ratio (PR) can be regarded as a continuous function of the percentile. In this case, we can consider the k-PR, the ratio of the survival distributions at the kth percentile, as one of a possible set of measures of the treatment effect. For illustration, we focus on AFT distributions defined in the extended log-gamma distribution, initially presented by Prentice in 9. For this family of distributions, the PR does not vary with the percentile chosen and is equal to the acceleration factor for the AFT models and can also be shown to be equal to the exponentiation of the treatment effect. These are also considered in combination with a PH model with log-logistic baseline, a model which does not have a constant PR. Other families of distributions could be considered (log F 9, log Burr 10), but our aim is simply to allow variation in the representation of PRs and a wide scope in the choice of parametric form. In particular, this allows distributional variation across studies which goes beyond that represented by random effects, or frailty, time-to-event models. These may be suitable for some multi-center trials or meta-analyses but typically only allow a random shift in one parameter across centers or studies. Note that while non-parametric estimation of percentiles is also possible, the generality of the parametric approach maintains considerable flexibility in distributional shapes while also enabling the incorporation of covariates into the meta-analysis in a natural manner. We begin in Section 2 by introducing a motivating example of the meta-analysis of glioma studies. In Section 3, we consider the PR as a measure of treatment efficacy. Maximum likelihood inference is also considered. In Section 4 a discussion of AFT models as well as the details of the extended log-gamma model are presented, while Section 5 gives details of how AFT models can be combined with log-logistic PH models in a meta-analysis framework. A discussion about study heterogeneity is presented in Section 6 followed by the analysis of glioma data in Section 7. The paper concludes with a discussion in Section 8.

2. Motivation: glioma example

We consider, as an example, an IPD meta-analysis of 12 randomized controlled trials investigating the use of chemotherapy in patients with high-grade glioma 11. Patients in the treatment groups were treated with surgery, radiotherapy and chemotherapy, while patients in the control groups were treated with surgery and radiotherapy. In the original report hazard ratios were estimated for each trial using log-rank analyses, and pooled hazard ratios were calculated for various outcomes using the fixed effects model. The overall hazard ratio for time to death was found to be 1.18 (95 per cent CI 1.09–1.28) comparing the treatment group with the control group. This was one of the meta-analyses undertaken at the MRC(UK) Clinical Trials Unit which prompted the methodological investigation reported here. The models we propose are alternatives to PH models. It is therefore interesting to examine the extent to which the PH assumption is violated in this data set. In Figure 1 for each trial the complementary log–log of the estimated survivor function has been plotted against the log of the survival times for the control and treatment groups, separately. Here the Kaplan–Meier estimate of the survivor function has been used. Under Cox's PH model the distance between the treatment and control curves in these graphs should be the same at all times. We can therefore assess by eye whether or not Cox's PH assumption is appropriate. From Figure 1 we conclude that the PH assumption is violated in some trials, the worst violation being in trial 17, while in others it appears appropriate, for example in trial 13.
Figure 1

Assessing violations of Cox's PH assumption in the trials from the glioma data set.

Assessing violations of Cox's PH assumption in the trials from the glioma data set. In Section 7 we will reanalyze the glioma data using parametric models. We will initially restrict our attention to AFT models. The advantage of this approach lies in its simplicity, because for AFT models the PR is constant across percentiles (see Section 4.1). We will make use of the extended log-gamma distribution, which will be described in detail in Section 4.2, because it incorporates several common parametric AFT models for survival data, including the Weibull and log-normal distributions. Within this family it is not necessary to make a choice about the appropriate distribution for each trial but choices of distributions outside this family are possible. In order to illustrate this subsequently, we will choose to use log-gamma models for trials 9, 11, 17 and 19 while using a PH model with log-logistic baseline for the remaining trials for which a PH assumption appears most appropriate.

3. Measuring treatment efficacy

3.1. The percentile ratio

The k-PR q can be defined as where k can take any value in [0, 1]. This quantity is thus relevant to any binary explanatory variable specifying group membership, such as a treatment identifier, and provides a relative measure for the treatment effect at each point on the survival probability axis. When discussing binary explanatory variables subsequently, we will assume that it is a treatment versus control comparison which is of interest. Note, however, that if a continuous explanatory variable is of interest, then q can be defined as the PR that reflects a unit change in the chosen variable. For k = 0.5, the quantity q0.5 represents the median ratio, possibly the percentile ratio of most general interest. Values >1 indicate, for example, that the median survival of the treatment group at this particular percentile is greater than the median survival of the control group, while values <1 indicate the opposite. In some circumstances, of course, another PR may be of more interest. In the most general setting, q changes as a function of k, since the PR for a specific value k does not capture the effect of treatment over the entire follow-up period of a trial. For that reason we might have to consider q over a range of values of k. For illustration, in Figure 2 the PR is plotted for a PH distribution with a log-logistic baseline, with (a) positive and (b) negative treatment effects. We consider percentiles only in the [0.05, 0.95] interval, since calculating q at the two limiting points, 0 and 1, is not informative. However, the limit for k→1 is calculable and can be taken to represent a final PR at the end of the study. At values of k close to zero, on the other hand, q is unstable, so conventionally we acknowledge that no treatment is better at k = 0, and hence q0 = 1. We thus adopt the notation k∈(0, 1) to imply that the extreme values 0 and 1 are not considered.
Figure 2

Plot of the PR for a proportional hazards distribution with log-logistic baseline, with positive (a) and negative (b) treatment effects.

Plot of the PR for a proportional hazards distribution with log-logistic baseline, with positive (a) and negative (b) treatment effects.

3.2. Likelihood inference

For now we focus on the inference concerning the PR q for a particular percentile level k. The situation when there is no natural or consensus choice of k, for inference purposes, is discussed later. Suppose we want to model the data from a study using a distribution f(t;v, β) for the time to an event T, where v is a parameter which characterizes the treatment effect, and β is a vector containing all other parameters relevant to the distribution. Irrespective of the choice of distribution f(), we can reparameterize it as f(t;q, β) by expressing v as a function of q, and possibly β, say v = g(q, β), conveniently written in this form to also highlight its dependence on the choice of k. However, q is a quantity with a clear interpretation and its scale does not depend on the choice of distribution f() or indeed any other features of the data being analyzed. Therefore, within a parametric meta-analysis, where different distributions are fit to data from different studies, q presents a measure of treatment effect in each of the separate analyses but remains directly comparable across studies. This means that, as a basis for meta-analysis, there exists a parameter common across distributions with an interpretation that can be easily communicated. Consider now the case where we have N studies to be pooled for a meta-analysis and where we assume that f(t;b, u|x) is the chosen distributional form to model the data in study i (i = 1, …, N), where j (j = 1, …, n) denotes the individuals in study i, b is a scale parameter and u = µ + vx is the location parameter represented as a function of explanatory variables x, denoting treatment and other relevant patient-specific information. Still focussing on a particular percentile level k, we can express the distribution for study i as f(t;µ, b, |x) using a reparameterization as discussed in the previous paragraph. Here is the k-PR of study i. The most common assumption in a meta-analysis, of IPD or otherwise, is that the true value of the quantity of interest is the same across studies, while other parameter values can vary. Therefore, we fix for all studies. Then, the likelihood function can be written as where I is the usual indicator variable for events. Also, the usual assumption that censoring is non-informative, in each of the N studies, is made. Based on (2), standard maximum likelihood estimation (MLE) of the common parameter q is possible. Previously in this section we have focussed on the inference concerning a particular value of the percentile level k. However, it may be more appropriate to consider a range of values of k. In this case we can carry out a separate analysis for each value of k and plot the results against k. Since the reparameterization procedure we used to derive the likelihood (2) may depend on k, for every choice of k there is a different likelihood. The likelihoods can only lead to identical inferences if the q's can be jointly modeled to be common across studies for every k, generally only true if q = qfor all k or if the dependence of q on k is modeled to be the same across studies through an assumption of a common distributional shape. The second possibility will only be true under restrictive assumptions about the common features of the time-to-event distributions across trials. The first is less restrictive in regard to distributional shape and is, for example, satisfied if estimation is based on the log-gamma family of AFT distributions. More generally, we suggest that , which is defined by maximization of (2) for a specific k, be regarded as a particular pooled summary of the that would be calculated from the studies individually. Then , viewed as a continuous function of k∈(0, 1), represents these pooled summaries and will provide some indication of the variation of the treatment effect over k. The values, for separate values of k, will be correlated but there appears to be no compelling practical reason to consider formal simultaneous inference that would depend on this correlation since separate 's are based on different model assumptions. For illustration, Figure 3 presents a simple case where, for five different studies, we have generated data from PH log-logistic distributions, as in the example in Figure 2. There was no censoring and the data were generated based on different sets of parameters, with the common characteristic q.5 = 2. Clearly q for values of k≠0.5 is not the same across studies, especially for small k. Based on the meta-analysis framework introduced in this section, we obtain the pooled estimate , plotted for k∈[0.05, 0.95], which nicely falls in the center of individual study curves that give the estimated q values and thus summarizes them in a single curve. For k = 0.5 we get , very close to the true value of 2.
Figure 3

Plot of data generated from PH log-logistic distributions for five studies with different parameter values but a common median percentile ratio q0.5 = 2. The pooled estimate has also been plotted.

Plot of data generated from PH log-logistic distributions for five studies with different parameter values but a common median percentile ratio q0.5 = 2. The pooled estimate has also been plotted.

4. Parametric models for meta-analysis

4.1. AFT models

A parametric AFT model for a random variable Y on (−infty, infty) corresponds to the use of a distribution with p.d.f. where u(−∞≤u≤∞) and b>0 are the location and scale parameters, respectively, f0(z) is a p.d.f on (−∞, ∞) and Y = log T, where T represents time. The distribution and survival functions for Y are F0[(y − u)/b] and S0[(y − u)/b] respectively, where The survival function for T = exp(Y) can be expressed as where α = exp(u), β = b−1 and is the survival function defined by the relationship . For treatment comparisons based on a treatment indicator x, we would let u = u(x) = µ + vx. For a more detailed discussion of log-location-scale models see 12. This model can also be expressed as a regression model where E is a random variable with p.d.f. f0(z). This has been a very useful model for the parametric analysis of time-to-event data. Extreme value, normal and logistic distributions for Y, correspond to Weibull, log-normal and log-logistic distributions for T, which are three of the most popular distributions for this purpose. For this model, if and are kth percentiles for the time-to-event distribution in treatment and control groups, respectively, then and thus q does not depend on k. This is a known and expected result for AFT models as the explanatory variables act multiplicatively on the time scale, and q is equal to the acceleration factor for all k. If every distribution f(t;b, u|x), as defined in Section 3.2, has this AFT structure, then the reparameterization as suggested in (4), is trivial and the assumption that PRs are constant across studies is simply an assumption that regression coefficients are the same across studies. The estimation of q, which equals a common value q for all k, can then be based on a single likelihood defined by (2). Another advantage of restricting attention to AFT models is that it is straightforward to include covariates, by adding extra terms to the right-hand side of equation (3). Within an IPD meta-analysis, patient-level covariates can be added in one of two ways. The covariate effects can be constrained to be the same in all studies or they can be allowed to vary across studies. In the former case the assumption of equal covariate effects can be tested by using a likelihood ratio test to compare the two models.

4.2. The extended log-gamma model

A useful class of AFT models is represented by the extended log-gamma model. This is a regression model of the form (3), where the error p.d.f. is written as which is a representation of Γ(κ)−1exp(κu − exp(u)), the log-gamma distribution, following some model manipulation and with w = (Y − µ − vx)/b. The survival function S is given by where r = γw − 2ln(γ), is the incomplete gamma ratio and is the incomplete normal integral. This distribution was initially discussed in 9, where an extension to the log-gamma model was considered by allowing γ<0, with the p.d.f. at −γ being a reflection about the origin of that at γ. Further discussion about this model can be found in 13, together with illustrative applications and additional references. A recent tutorial paper 14 on the use of the generalized gamma distribution for survival analysis has also appeared. This three-parameter family of distributions, where the parameter γ, together with b, specifies the form of the error density function, includes the special cases for T = e of Weibull (γ = 1), exponential (γ = b = 1), log-normal (γ = 0), gamma (γ = b), generalized gamma (γ>0) and reciprocal Weibull (γ = − 1) distributions. More generally, through estimation of γ, we minimize the need for assumptions about error distributions in various studies being considered. Of course, there is a need to estimate the N γ parameters but this should not be a problem if each study is of moderate size as evidenced by the examples in 13. In an electronic appendix‡ we investigate the empirical behavior of meta-analyses based on PRs for the special case of the log-gamma family of distributions and demonstrate that it performs as expected. We compare the extended log-gamma model to models which assume all Weibull, log-normal or log-logistic distributions, and find that the PR standard errors are smaller for the extended log-gamma model because it provides a better fit to the data through allowing the underlying time-to-event distributions to vary across studies. Since Weibull distributions are linked to PH models of widely varying shapes and retain the AFT assumption of common PRs, the log-gamma family was convenient for our simulation study and, more generally, may represent a useful approach to the meta-analysis of multiple trials where not all will necessarily be consistent with a PH representation of a treatment effect. However, to illustrate that this is not a necessary restriction we consider an alternative PH family in the following section.

5. Incorporating log-logistic PH models

In this section, we consider the more complicated situation when AFT models are appropriate for some studies whereas PH models which are not also in the AFT class are appropriate for others. For this purpose, we can, as was done to produce Figure 2, consider a PH model h(t;θ|x) = eθh0(t) with log-logistic baseline function, where are the baseline hazard and survival functions, respectively. It can then be easily shown that which makes explicit that the relationship between θ, the logHR, and the PR is a function of k. If (5) is used to reparameterize this PH model, then we can use both AFT and PH models in the likelihood (2) to make inferences concerning q. Because the inclusion of PH models means that q will depend on k, we could, as suggested earlier, plot against k or focus on a particular value of k of interest.

6. Heterogeneity

Exploring heterogeneity is always advisable in meta-analysis. It is very important to feel confident that summary inferences drawn from multiple studies are informative. In RCTs where only aggregate data are available it is natural to obtain a forest plot and observe whether the estimates of the treatment effects across trials are similar. The parametric scenario, presented in this paper, provides a useful structure in order to test whether PRs are the same across studies. Along with the production of a forest plot, we can easily test the null hypothesis against the alternative of arbitrary differences based on a likelihood ratio test with N − 1 degrees of freedom. This test examines heterogeneity at the k-percentile, giving some information about the homogeneity of the studies. As for the summary measures themselves, when PRs vary with k and multiple tests can be performed, formal simultaneous inference concerning heterogeneity is unlikely to be of central interest.

7. Analysis of glioma example

In this section we return to the example described in Section 2. We were unable to obtain permission to use the data from 3 of the 12 trials in the original meta-analysis. We will therefore use data from nine trials in our analysis. Model fitting used the R software for statistical analysis (http://www.r-project.org), in which code was written to generate and analyze the data (available in the electronic appendix). Of readily available alternative software, STATA may be an attractive alternative since its parametric survival regression package includes the extended log-gamma distribution as one of the default distributions. Furthermore, it allows all ancillary parameters to be estimated separately in each study while restricting regression coefficients to be the same across studies. However, it does fit the same distribution for all studies and would also restrict regression coefficients other than that associated with treatment to be the same across studies. These are possible drawbacks. We analyze the data initially using AFT models. We use the extended log-gamma distribution, taking advantage of its flexible nature by allowing the shape parameter to vary between trials. The location and scale parameters of the log-gamma distribution are also allowed to vary between trials, while the logarithm of the percentile ratio (logPR) is assumed to be the same in all trials. We obtain parameter estimates by maximizing the likelihood (2) over all parameters simultaneously, using optim in R and their asymptotic variance–covariance matrix was estimated by the inverse of the observed information matrix. The pooled logPR estimate for the glioma data is 0.163 (95 per cent CI 0.071, 0.255), which corresponds to a PR of 1.176. Recall that for AFT models the PR is constant across percentiles. The PR can therefore be explained as the ratio of the survival time for a patient on the experimental treatment to the survival time for a patient on the control treatment for any percentile. A profile likelihood plot for the logPR is shown in Figure 4 to demonstrate that the likelihood is symmetric about the estimated logPR. In Figure 4 a constant has been added to the loglikelihood so that the maximized value is 1.92; confidence intervals are therefore given by the intersection of the curve with the x-axis.
Figure 4

Profile log-likelihood plot for the logPR for the glioma data analysis using the extended log-gamma model.

Profile log-likelihood plot for the logPR for the glioma data analysis using the extended log-gamma model. A forest plot for the glioma data based on the extended log-gamma model is shown in Figure 5, along with estimates of the logPRs and shape parameters from each trial. In this plot the individual trial estimates derive from an extended log-gamma analysis of each study individually. The test for heterogeneity described in Section 6 gives a likelihood ratio statistic of 11.4 for the glioma data, which corresponds to a p-value of 0.18 when compared with a chi-squared distribution with eight degrees of freedom.
Figure 5

Forest plot for the ELG meta-analysis of the glioma data. The table gives the estimated logPR for each trial, along with the estimated value of the shape parameter for each trial.

Forest plot for the ELG meta-analysis of the glioma data. The table gives the estimated logPR for each trial, along with the estimated value of the shape parameter for each trial. We demonstrate the addition of patient-level covariates here by adjusting the model for sex. As described in Section 4.1, this can either be done by constraining the covariate effect to be the same in all trials, or by allowing it to vary between trials. The first method, with identical covariate effects in all trials, gives an estimated logPR of 0.20 (95 per cent CI 0.12, 0.28), which is slightly higher than the unadjusted estimate of 0.16. Allowing covariate effects to vary between trials, the estimated logPR is 0.19 (95 per cent CI 0.11, 0.27). The two methods therefore give similar results in this case, although a likelihood ratio test comparing the two gives a chi-squared statistic of 18.49 on eight degrees of freedom, which corresponds to a p-value of 0.018. This suggests that use of the more complicated model is justified here. Other covariates which may have been interesting to include in this example are histology and the extent of resection, but these have been omitted due to the presence of missing data. In the second analysis of the same data set we use a combination of PH models with log-logistic baselines and AFT models. In Figure 1 we presented some diagnostic plots which can be used to assess whether or not a PH assumption is appropriate for the data from each trial. On the basis of these plots we model the data from trials 9, 11, 17 and 19 using AFT distributions and the data from the remaining trials using log-logistic PH distributions. As before we will take the PR to be constant across trials. However, because the PR may now vary across percentiles, we must impose this assumption separately for each percentile. This means that we are effectively fitting a different model at each percentile. We maximize the likelihood (2) for each percentile using the extended log-gamma distribution for data from the first set of trials, and the log-logistic PH distribution, described in Section 5, for data from the remaining trials. For the PH distributions, equation (5) is used to convert the logHR into the logPR. The results from fitting the PH/AFT model are presented in the graph in Figure 6. Estimates for the pooled logPR are plotted as a solid line, with 95 per cent confidence intervals as dashed lines. The logPR was estimated for percentiles at intervals of 0.1, indicated by circles on the curves in the graph. The estimated logPR decreases as the percentile increases, which is a consequence of using the log-logistic hazard function as a baseline in the PH models. Note that the curve is fairly flat, which means that an analysis assuming AFT models for all trials may be justified in this case. Also plotted as a dotted line in Figure 6 is the estimated logPR from the first analysis. This result lies in the middle of the PH/AFT results, coinciding with the PH/AFT curve around the median percentile ratio.
Figure 6

Results of the PH/AFT analysis—estimated logPR's are plotted as a solid line and 95 per cent confidence intervals around these estimates are plotted as dotted lines. The circles indicate the percentiles at which logPR's were estimated. The estimated logPR from the analysis which assumes AFT models for all trials is plotted as a dotted line.

Results of the PH/AFT analysis—estimated logPR's are plotted as a solid line and 95 per cent confidence intervals around these estimates are plotted as dotted lines. The circles indicate the percentiles at which logPR's were estimated. The estimated logPR from the analysis which assumes AFT models for all trials is plotted as a dotted line. Our results are in keeping with the finding of the original meta-analysis of this data set, which however used data from all 12 trials. The original result was a pooled hazard ratio of 1.18 comparing radiotherapy plus chemotherapy to radiotherapy alone. This corresponds to a logHR of 0.16, which is very close to our pooled AFT logPR estimate of 0.16. The PR estimate has the advantage that its meaning is easier to interpret for practitioners, being the estimated increase in survival time for those in the experimental group compared with those in the control group.

8. Discussion

In this paper, our aim was to discuss the potential value of percentile ratios in the IPD meta-analysis of time-to-event outcomes. Because percentile ratios can be defined for the comparison of any two survival curves, they provide a reasonable basis on which to compare and combine treatment effects across studies. For this reason, we feel they are worth investigation as a basis for IPD meta-analysis of time-to-event outcomes. Ultimately their advantages and disadvantages with respect to analyses based on hazard ratios can only emerge as they are used in a variety of settings. When each study is modeled using the correct distribution and the PR of interest, q, is common across studies, then MLE will provide unbiased estimates for this PR. Even if percentile ratios from different studies, the 's, vary then will still be a potentially useful pooled summary of these values. Further work should explore the consequences of model misspecification more generally however. Different time-to-event models may be adopted for different studies, but the extended log-gamma model represents a broad class of distributions which may be particularly useful. An AFT family of distributions has the particular advantage that the PR is invariant across the percentile level. An analysis based on extended log-gamma models is therefore simple and relatively undemanding computationally, and provides a single estimate of treatment effect which is easily interpretable. Alternatively, distributions may be considered for which the PR does depend on the percentile level. Then it may be that a single or a set of percentile levels is of interest. The choice is likely to be context specific. Here, we have focussed on the use of PH distributions with log-logistic baseline hazard functions in combination with AFT distributions. In practice, however, alternative distributions for the baseline hazard could be considered. This is a more flexible approach than assuming AFT distributions for all studies, but more work is required to investigate how the possibly subjective choice of distribution for each study might affect the results. Here, we have used graphical methods to assess departure from the PH assumption, but with a larger data set more sophisticated methods such as those proposed by Boutitie et al. 15 might be used. Subsequently, our aim is to investigate random effects models in this context in which the PRs from different studies are assumed to come from a known distribution. A possible hierarchical regression model, for this purpose, could introduce random effects v as follows: v = log (q) is considered to be the average log-PR and g is the deviation from the log-PR in study i. Thus the q's are no longer assumed to be the same across studies but instead are assumed to come from a Normal distribution with mean v and variance τ2. The WinBugs software can be used to fit random effects models with known distributions for each study, and these need not be AFT. However, the extended log-gamma model is not available by default in WinBugs and so separate development in this package or otherwise will be required.
Table AI

Set of parameters for simulations with the extended log-gamma model.

γμβ
Sim A1(0.3, 0.6, 0.9, 1.2, 1.5)(7,7,7,7,7)(1,1,1,1,1)
Sim A2(0.3, 0.6, 0.9, 1.2, 1.5)(4,9,7,3,8)(1,1,1,1,1)
Sim A3(0.3, 0.6, 0.9, 1.2, 1.5)(4,9,7,3,8)(1.5,0.6,1.2,0.8,1.1)
Sim B1(−2, −1, 0.3, 1, 2)(7,7,7,7,7)(1,1,1,1,1)
Sim B2(−2, −1, 0.3, 1, 2)(4,9,7,3,8)(1,1,1,1,1)
Sim B3(−2, −1, 0.3, 1, 2)(4,9,7,3,8)(1.5,0.6,1.2,0.8,1.1)
Table AII

Results of the A simulations, based on various levels of censoring. Estimates are given with standard deviations of estimates in curved brackets and mean estimated standard errors in square brackets.

Stratified analysisSingle study analysis


WeibLLLNWeibLLLNELG
Sim A1 (0 per cent)0.69690.69760.69770.69650.69740.69770.6971
(0.0643)(0.0749)(0.0817)(0.0654)(0.0741)(0.0817)(0.0628)
[0.0640][0.0759][0.0820][0.0652][0.0768][0.0843][0.0626]
Sim A1 (20 per cent)0.69030.69010.68950.69040.69030.68940.6916
(0.0745)(0.0848)(0.0950)(0.0748)(0.0840)(0.0956)(0.0748)
[0.0711][0.0827][0.0904][0.0724][0.0840][0.0931][0.0693]
Sim A1 (40 per cent)0.68840.68710.68760.68800.68730.68770.6894
(0.0848)(0.0929)(0.1062)(0.0850)(0.0921)(0.1079)(0.0834)
[0.0817][0.0928][0.1028][0.0832][0.0947][0.1065][0.0791]
Sim A2 (0 per cent)0.69470.69510.69600.69470.69700.69600.6948
(0.0676)(0.0794)(0.0863)(0.0897)(0.0907)(0.0863)(0.0655)
[0.0640][0.0757][0.0818][0.1440][0.1774][0.1670][0.0626]
Sim A2 (20 per cent)0.69910.69820.69660.69640.69910.69710.6987
(0.0768)(0.0859)(0.0953)(0.1127)(0.1287)(0.1213)(0.0753)
[0.0711][0.0827][0.0904][0.1612][0.1946][0.1864][0.0694]
Sim A2 (40 per cent)0.69810.69770.69910.69430.69880.69830.6965
(0.0838)(0.0945)(0.1062)(0.1492)(0.1689)(0.1644)(0.0846)
[0.0818][0.0929][0.1031][0.1867][0.2189][0.2148][0.0791]
sim A3 (0%)0.69810.70100.70130.69950.70220.70130.6961
(0.0753)(0.0757)(0.0830)(0.0687)(0.0950)(0.0830)(0.0564)
[0.0687][0.0785][0.0870][0.1422][0.1829][0.1688][0.0557]
sim A3 (20 per cent)0.69250.69480.69650.69840.70160.70120.6926
(0.0763)(0.0846)(0.0966)(0.1030)(0.1327)(0.1216)(0.0638)
[0.0746][0.0854][0.0954][0.1612][0.2011][0.1888][0.0617]
sim A3 (40 per cent)0.69460.69550.69540.69260.69420.69260.6927
(0.0825)(0.0888)(0.1040)(0.1384)(0.1704)(0.1590)(0.0697)
[0.0852][0.0964][0.1084][0.1904][0.2281][0.2190][0.0707]
Table AIII

Results of the B simulations, based on various levels of censoring. Estimates are given with standard deviations of estimates in curved brackets and mean estimated standard errors in square brackets.

Stratified analysisSingle study analysis


WeibLLLNWeibLLLNELG
Sim B1* (0 per cent)0.63430.70030.69100.60090.69960.69100.6972
(0.1070)(0.0816)(0.0913)(0.1512)(0.0724)(0.0913)(0.0630)
[0.0878][0.0870][0.0955][0.1170][0.0950][0.1125][0.0626]
Sim B1 (20 per cent)0.67390.69180.68920.65640.69220.68840.6926
(0.0989)(0.0896)(0.1019)(0.1396)(0.0848)(0.1059)(0.0705)
[0.0874][0.0912][0.1007][0.1134][0.1026][0.1200][0.0668]
Sim B1 (40 per cent)0.69660.69450.69480.69400.69350.69320.6925
(0.0926)(0.0911)(0.1061)(0.1133)(0.0938)(0.1169)(0.0740)
[0.0911][0.0977][0.1102][0.1136][0.1135][0.1332][0.0733]
Sim B2 (0 per cent)0.64050.69200.68770.64690.69430.68770.6960
(0.1404)(0.0793)(0.0924)(0.1161)(0.1001)(0.0924)(0.0629)
[0.0986][0.0876][0.0983][0.1591][0.1819][0.1758][0.0623]
Sim B2 (20 per cent)0.69790.69620.69870.69610.70200.70050.6963
(0.1010)(0.0895)(0.1021)(0.1374)(0.1283)(0.1259)(0.0715)
[0.0881][0.0916][0.1012][0.1750][0.1973][0.1936][0.0670]
Sim B2 (40 per cent)0.69130.69100.69210.68140.68930.68670.6963
(0.0952)(0.0963)(0.1101)(0.1613)(0.1681)(0.1658)(0.0761)
[0.0911][0.0975][0.1094][0.1984][0.2220][0.2204][0.0734]
Sim B3* (0 per cent)0.62150.68730.67850.64170.68580.67850.6902
(0.1217)(0.0887)(0.1065)(0.1249)(0.1179)(0.1065)(0.0600)
[0.1011][0.0934][0.1094][0.1553][0.1859][0.1762][0.0558]
Sim B3 (20 per cent)0.68840.69340.69220.68670.68880.68880.6944
(0.1103)(0.0918)(0.1106)(0.1256)(0.1411)(0.1334)(0.0644)
[0.0954][0.0983][0.1116][0.1648][0.1985][0.1913][0.0594]
Sim B3 (40 per cent)0.69870.69940.70270.70420.70430.70470.6983
(0.0937)(0.0946)(0.1127)(0.1371)(0.1632)(0.1577)(0.0665)
[0.0949][0.1025][0.1168][0.1890][0.2201][0.2154][0.0651]

Event times have been truncated at 250 000.

Table AIV

Coverage probabilities.

Stratified analysisSingle study analysis


WeibLLLNWeibLLLNELG
Sim A1 (0 per cent)0.9520.9440.9420.9460.9520.9420.952
Sim A1 (20 per cent)0.9340.9420.9340.9400.9440.9380.928
Sim A1 (40 per cent)0.9460.9520.9380.9440.9620.9380.944
Sim A2 (0 per cent)0.9300.9380.9340.9980.9980.9980.930
Sim A2 (20 per cent)0.9260.9400.9340.9961.0001.0000.930
Sim A2 (40 per cent)0.9520.9400.9440.9900.9940.9940.932
Sim A3 (0 per cent)0.9260.9480.9521.0001.0001.0000.946
Sim A3 (20 per cent)0.9440.9560.9561.0000.9980.9980.940
Sim A3 (40 per cent)0.9440.9600.9520.9980.9900.9920.958
Sim B1 (0 per cent)0.8520.9520.9600.7920.9880.9880.948
Sim B1 (20 per cent)0.9240.9560.9480.8900.9740.9740.938
Sim B1 (40 per cent)0.9300.9640.9500.9400.9860.9780.942
Sim B2 (0 per cent)0.7980.9680.9600.9841.0001.0000.928
Sim B2 (20 per cent)0.9160.9600.9520.9921.0001.0000.938
Sim B2 (40 per cent)0.9400.9540.9420.9861.0001.0000.938
Sim B3 (0 per cent)0.8420.9520.9540.9720.9981.0000.926
Sim B3 (20 per cent)0.9140.9580.9480.9920.9960.9920.924
Sim B3 (40 per cent)0.9460.9540.9740.9960.9900.9980.924
  10 in total

1.  The utility of mixed-form likelihoods.

Authors:  R J Cook; V T Farewell
Journal:  Biometrics       Date:  1999-03       Impact factor: 2.571

2.  Aggregate data meta-analysis with time-to-event outcomes.

Authors:  Paula R Williamson; Catrin Tudur Smith; Jane L Hutton; Anthony G Marson
Journal:  Stat Med       Date:  2002-11-30       Impact factor: 2.373

3.  Parametric survival analysis and taxonomy of hazard functions for the generalized gamma distribution.

Authors:  Christopher Cox; Haitao Chu; Michael F Schneider; Alvaro Muñoz
Journal:  Stat Med       Date:  2007-10-15       Impact factor: 2.373

4.  Meta-analysis of pairs of survival curves under heterogeneity: a Poisson correlated gamma-frailty approach.

Authors:  M Fiocco; H Putter; J C van Houwelingen
Journal:  Stat Med       Date:  2009-12-30       Impact factor: 2.373

5.  Extracting summary statistics to perform meta-analyses of the published literature for survival endpoints.

Authors:  M K Parmar; V Torri; L Stewart
Journal:  Stat Med       Date:  1998-12-30       Impact factor: 2.373

6.  Assessing treatment-time interaction in clinical trials with time to event data: a meta-analysis of hypertension trials.

Authors:  F Boutitie; F Gueyffier; S J Pocock; J P Boissel
Journal:  Stat Med       Date:  1998-12-30       Impact factor: 2.373

7.  Meta-analysis of individual patient data from randomized trials: a review of methods used in practice.

Authors:  Mark C Simmonds; Julian P T Higgins; Lesley A Stewart; Jayne F Tierney; Mike J Clarke; Simon G Thompson
Journal:  Clin Trials       Date:  2005       Impact factor: 2.486

8.  Investigating heterogeneity in an individual patient data meta-analysis of time to event outcomes.

Authors:  Catrin Tudur Smith; Paula R Williamson; Anthony G Marson
Journal:  Stat Med       Date:  2005-05-15       Impact factor: 2.373

9.  To IPD or not to IPD? Advantages and disadvantages of systematic reviews using individual patient data.

Authors:  Lesley A Stewart; Jayne F Tierney
Journal:  Eval Health Prof       Date:  2002-03       Impact factor: 2.651

Review 10.  Chemotherapy in adult high-grade glioma: a systematic review and meta-analysis of individual patient data from 12 randomised trials.

Authors:  L A Stewart
Journal:  Lancet       Date:  2002-03-23       Impact factor: 79.321

  10 in total
  12 in total

1.  Meta-analysis of time-to-event outcomes from randomized trials using restricted mean survival time: application to individual participant data.

Authors:  Yinghui Wei; Patrick Royston; Jayne F Tierney; Mahesh K B Parmar
Journal:  Stat Med       Date:  2015-06-23       Impact factor: 2.373

2.  Two-stage meta-analysis of survival data from individual participants using percentile ratios.

Authors:  Jessica K Barrett; Vern T Farewell; Fotios Siannis; Jayne Tierney; Julian P T Higgins
Journal:  Stat Med       Date:  2012-07-24       Impact factor: 2.373

3.  Individual patient data meta-analysis of survival data using Poisson regression models.

Authors:  Michael J Crowther; Richard D Riley; Jan A Staessen; Jiguang Wang; Francois Gueyffier; Paul C Lambert
Journal:  BMC Med Res Methodol       Date:  2012-03-23       Impact factor: 4.615

4.  A multivariate model for the meta-analysis of study level survival data at multiple times.

Authors:  Dan Jackson; Katie Rollins; Patrick Coughlin
Journal:  Res Synth Methods       Date:  2014-02-27       Impact factor: 5.273

5.  Dose and Fractionation in Radiation Therapy of Curative Intent for Non-Small Cell Lung Cancer: Meta-Analysis of Randomized Trials.

Authors:  Johanna Ramroth; David J Cutter; Sarah C Darby; Geoff S Higgins; Paul McGale; Mike Partridge; Carolyn W Taylor
Journal:  Int J Radiat Oncol Biol Phys       Date:  2016-07-25       Impact factor: 7.038

6.  Comparison of Partial Hepatectomy and Transarterial Chemoembolization in Intermediate-Stage Hepatocellular Carcinoma: A Systematic Review and Meta-Analysis.

Authors:  Ismail Labgaa; Patrick Taffé; David Martin; Daniel Clerc; Myron Schwartz; Norihiro Kokudo; Alban Denys; Nermin Halkic; Nicolas Demartines; Emmanuel Melloul
Journal:  Liver Cancer       Date:  2020-01-28       Impact factor: 11.740

7.  Bias and precision of methods for estimating the difference in restricted mean survival time from an individual patient data meta-analysis.

Authors:  Béranger Lueza; Federico Rotolo; Julia Bonastre; Jean-Pierre Pignon; Stefan Michiels
Journal:  BMC Med Res Methodol       Date:  2016-03-29       Impact factor: 4.615

Review 8.  Get real in individual participant data (IPD) meta-analysis: a review of the methodology.

Authors:  Thomas P A Debray; Karel G M Moons; Gert van Valkenhoef; Orestis Efthimiou; Noemi Hummel; Rolf H H Groenwold; Johannes B Reitsma
Journal:  Res Synth Methods       Date:  2015-08-19       Impact factor: 5.273

9.  Less is more in solid-dominant lung cancer? Sublobar resection versus lobectomy for solid-dominant stage IA non-small-cell lung cancer: A meta-analysis study.

Authors:  Juntang Guo; Yang Liu; Xiaodong Tian; Zhipeng Ren; Jixing Lin; Bailin Wang; Chaoyang Liang
Journal:  Mol Clin Oncol       Date:  2019-08-22

Review 10.  Individual participant data meta-analysis of intervention studies with time-to-event outcomes: A review of the methodology and an applied example.

Authors:  Valentijn M T de Jong; Karel G M Moons; Richard D Riley; Catrin Tudur Smith; Anthony G Marson; Marinus J C Eijkemans; Thomas P A Debray
Journal:  Res Synth Methods       Date:  2020-02-06       Impact factor: 5.273

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.