Literature DB >> 30073685

Blinded and unblinded sample size reestimation procedures for stepped-wedge cluster randomized trials.

Michael J Grayling1, Adrian P Mander1, James M S Wason1,2.   

Abstract

The ability to accurately estimate the sample size required by a stepped-wedge (SW) cluster randomized trial (CRT) routinely depends upon the specification of several nuisance parameters. If these parameters are misspecified, the trial could be overpowered, leading to increased cost, or underpowered, enhancing the likelihood of a false negative. We address this issue here for cross-sectional SW-CRTs, analyzed with a particular linear-mixed model, by proposing methods for blinded and unblinded sample size reestimation (SSRE). First, blinded estimators for the variance parameters of a SW-CRT analyzed using the Hussey and Hughes model are derived. Following this, procedures for blinded and unblinded SSRE after any time period in a SW-CRT are detailed. The performance of these procedures is then examined and contrasted using two example trial design scenarios. We find that if the two key variance parameters were underspecified by 50%, the SSRE procedures were able to increase power over the conventional SW-CRT design by up to 41%, resulting in an empirical power above the desired level. Thus, though there are practical issues to consider, the performance of the procedures means researchers should consider incorporating SSRE in to future SW-CRTs.
© 2018 The Authors. Biometrical Journal Published by WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

Entities:  

Keywords:  blinded; cluster randomized trial; internal pilot; sample size re-estimation; stepped-wedge

Mesh:

Year:  2018        PMID: 30073685      PMCID: PMC6175439          DOI: 10.1002/bimj.201700125

Source DB:  PubMed          Journal:  Biom J        ISSN: 0323-3847            Impact factor:   2.207


INTRODUCTION

A stepped‐wedge (SW) cluster randomised trial (CRT) involves the sequential roll‐out of an intervention across several clusters over multiple time periods, with the time period in which a cluster begins receiving the intervention determined at random. Recent papers have established methods for sample size determination in the case of cross‐sectional (Hussey & Hughes, 2007) and cohort (Hooper, Teerenstra, de Hoop, & Eldridge, 2016) designs, for trials with multiple levels of clustering and for incomplete block SW‐CRTs (Hemming, Lilford, & Girling, 2015). Undeniably, there has been a growing interest in the design, and in particular, it has now become associated with scenarios in which there is a belief that the trial's experimental intervention will be effective (Brown & Lilford, 2006; Mdege, Man, Taylor (nee Brown), & Torgerson, 2011). Given this commonly held belief, it may come as a surprise that a recent literature review determined that in 31% of the SW‐CRTs completed by February 2015, there was no significant effect of the experimental intervention on any of the trials primary outcome measures (Grayling, Wason, & Mander, 2017a). To guard against this, implicitly assuming this failure rate was due to the experimental interventions being futile, methodology for the incorporation of interim analyses in SW‐CRTs was recently described (Grayling, Wason, & Mander, 2017b). One other possible explanation is that the studies have been false negatives. A high false‐negative rate could be associated with SW‐CRTs having been underpowered. Methodology available to determine the sample size required by SW‐CRTs is dependent upon the specification of the values of several nuisance parameters (e.g., the between cluster and residual variances). In practice, it may be difficult to provide accurate estimates for these factors, and their misspecification may be leading to under‐powered studies. Alternatively, if these parameters are being misspecified such that SW‐CRTs have been overpowered, there may have been more measurements taken than actually required, leading to unnecessary cost. A common approach to addressing the specification of nuisance parameters in the trial design literature is the use of a sample size reestimation (SSRE) procedure. Each such method has essentially the same intention: to alleviate the issue of prespecifying nuisance parameters by allowing them to be reestimated during the trial, and the required sample size adjusted. Reviews of SSRE methodology have been provided by Friede and Kieser (2013), Proschan (2009), and Pritchett et al. (2015), among others. We refer the reader to these articles for a wider overview of SSRE. Of importance here is that, broadly speaking, SSRE procedures can be subcategorized according to whether they are a blinded or unblinded technique, and also in relation to whether they reestimate the variance parameters, the effect‐size, or some combination of the two. Regulatory agencies prefer that the blind be maintained whenever possible, so as to not risk compromising the validity of the trial (ICH, 1998). Fortunately therefore, blinded SSRE methodology is today available for a range of settings (e.g. Friede & Kieser, 2013; Golkowski, Friede, & Kieser, 2014; and Kunz, Stallard, Parsons, Todd, & Friede, 2017), with each such procedure typically conferring highly desirable trial operating characteristics. However, while some results exist on SSRE in multicentre (Jensen & Kieser, 2010) and parallel group CRTs (van Schie & Moerbeek, 2014), no work has established methodology for SSRE in SW‐CRTs, with the increased complexity in the design of SW‐CRTs necessitating a specialised approach. In this article, we address this by developing and exploring the performance of both blinded and unblinded SSRE procedures for cross‐sectional SW‐CRTs. Precisely, we consider the case in which the variance parameters required for sample size determination are to be reestimated, but the effect‐size is prespecified and remains unmodified. Accordingly, we assume that a commonly considered linear‐mixed model will be utilized for data analysis, and develop blinded estimators of the associated key variance parameters. The performance of a SSRE procedure based on these blinded estimators is then compared to an unblinded approach, as a function of their various control parameters, and the parameters of the underlying model. We conclude with a discussion of possible extensions to our approach, and by detailing logistical factors that must be considered when incorporating SSRE in to SW‐CRTs.

METHODS

Notation, hypotheses, and analysis

We suppose that a cross‐sectional SW‐CRT is to be carried out in C clusters over T time periods, with n individuals recruited per cluster per time period. That is, we assume data will be accrued on new patients in each cluster in each time period. We do not restrict our attention to “balanced” SW‐CRTs however; clusters need not start in the control condition, conclude in the experimental condition, and there does not need to be an equal number of clusters switching to the experimental intervention in each time period. We assume that the accumulated data will be normally distributed, and the following linear‐mixed model will be utilized for data analysis, as proposed by Hussey and Hughes (2007) Here is the response of the k‐th individual (), in the i‐th cluster (), in the j‐th time period (); μ is an intercept term; is a fixed effect for the j‐th time period (with for identifiability); τ is a fixed treatment effect for the experimental intervention relative to the control; is the binary treatment indicator for the i‐th cluster and j‐th time period. That is, if cluster i receives the intervention in time period j. We denote by X, , the matrix formed from the . Similarly, we define , , to be the matrix formed from the first t columns of X. That is, for and ; is a random effect for cluster i; is the individual‐level error. We denote the vector of fixed effects by . Moreover, we indicate the design matrix linking to the vector of responses , from the set of time periods , given an allocation matrix X, and a per cluster per period sample size of n, by . We similarly denote the covariance matrix of , given and , by . As is noted in Hussey and Hughes (2007), by the above choice of linear‐mixed model, is a block diagonal matrix. Each block, corresponding to the responses from each cluster, is of dimension . Precisely, it is given by , where and are the identity and unit matrices, respectively. We perform a one‐sided hypothesis test for τ and assume that it is desired to have a type‐I error rate of α when , and to have power to reject H 0 of when , for some specified . Note that SSRE procedures for two‐sided hypotheses are also easily achievable. Finally, we assume that hypothesized values for the variance parameters and have been provided, which we denote by and . Alternatively, a value for one of these parameters, and a value for the intracluster correlation (ICC) ρ, , could be specified, such that and can still be determined. Given these values, we assume a sample size calculation has been performed (using the methods to be described shortly) and values for X and n (and thus also C and T since ) have subsequently been specified. For reasons to be elucidated below, we refer to this n as . With the above, a conventional SW‐CRT can be conducted as follows. We recruit individuals per cluster per time period, with the experimental intervention allocated according to the matrix X. On completion, we use restricted error maximum likelihood (REML) estimation to acquire an estimate of , denoted , and a value for . Next, we compute the test statistic , and reject H 0 if , where e is the solution to Here, is the probability density function of a t‐distribution with mean μ, covariance Λ, and degrees of freedom ν, evaluated at x. Specifically, we take ν to be the degrees of freedom in a corresponding balanced multilevel ANOVA design. Later, we will discuss the implications of this, and describe other possible ways to prescribe ν. We next detail how the above can be extended to allow SSRE to be incorporated in to the design.

Sample size reestimation procedures

A single interim analysis for SSRE is included in the SW‐CRT design after a designated time period . Precisely, we assume that the trial is conducted as per matrix X and the value for time periods . After this, we compute estimates for the variance parameters, and , based upon the accumulated data. Below, we detail how exactly this is achieved in the blinded and unblinded procedures. Here, we discuss how these estimates are then used. Note that, as was discussed in Section 1, we are reestimating only and : the effect size δ will not be reestimated. We search numerically as follows to determine the required per cluster per period sample size for the remainder of the trial, , to convey the desired power if and . Thus, the number of clusters remains fixed throughout the trial; it is the per cluster per period sample size that is adjusted. We consider possible alternatives to this in the discussion. Firstly, suppose has been chosen, then time periods of the trial are conducted using the matrix X for treatment allocation, and recruiting individuals per cluster per period. At the end of the trial the linear‐mixed model (1) with REML estimation are utilized to acquire and as above. The test statistic is again determined, and H 0 rejected if , but e is now the solution to Here, is the degrees of freedom in a balanced multilevel ANOVA design if a sample size of is used per cluster per period in time periods , and a sample size of is used per cluster per period in time periods . The power to reject H 0 when , for a particular , can thus be estimated at the interim as where I is given by the inverse of element of the following matrix This matrix arises as the theoretical covariance matrix of the maximum likelihood estimator of when a sample size of is used per cluster per period in time periods , and a sample size of is used per cluster per period in time periods (see, e.g. Fitzmaurice, Laird, & Ware (2011) for details on how this form of covariance matrix is constructed). Therefore, we can compute the required value for by searching for the minimal integer solution to the following equation In fact, to make our SSRE procedures more applicable in practice, and to guard against unrealistically large values for , we carry out the remaining periods of the trial recruiting individuals per cluster per time period, where Here, and , with , are designated values for the minimal and maximal allowed number of patients per cluster per period following the reestimation. There are various possible approaches to how values for and can be specified. Wittes and Brittain (1990) advocated for the initially planned sample size to never be reduced. Here, this would correspond to . In contrast, Birkett and Day (1994) proposed that the initially planned sample size should be reduced. Gould (1992) recommended that a realistic maximum sample size should be specified. In practice, the values of and may need to be chosen based upon the potentially attainable values of n in a particular trial scenario. Finally, following determination of , the remainder of the trial and ensuant analysis is conducted as described above, to determine whether to reject H 0. Note that the sample size required by a classical fixed sample SW‐CRT design, given an allocation matrix X, can be determined using the above by treating as a variable rather than a fixed parameter, and searching for the minimal such that when . Alternatively, could be specified and the matrix X determined for the desired power. All that remains to be elucidated in the above procedure is the means of determining the estimates and . As discussed, we describe both blinded and unblinded approaches to their specification. The unblinded procedure is as follows. After time period t, we fit the following model to the accumulated data using REML estimation Here, , with . In the above, is included as a qualifier to indicate the term should appear in our model as at least one cluster has been administered the experimental intervention in some time period. Similarly, indicates period effects should be accounted for in the model. Following REML estimation, we attain our values for and immediately, and use them in the above algorithm to determine . For the blinded procedure, it is first useful to note precisely what we mean by “blinded.” Here, it refers to the fact that the treatment indicator of an observation is not known, but that it is known which observations belong to the sample cluster and time period. With this, define where In the Supplementary Material we derive that Given a particular choice for τ in the above, which we shall denote , these equations are used to estimate and as follows Compute and using the formulae above and the accrued data. Define as Set and Here, our specification of allows us to make use of when it will give rise to a value of such that , but assists against overcorrection for a nonzero treatment effect by specifying when . We then utilize the algorithm from earlier to determine the value of . Note that if , the above are unbiased estimators for the variance parameters and . For further clarity, the full unblinded and blinded SSRE procedures are detailed algorithmically in the Supplementary Material. In addition, we also discuss in the Supplementary Material possible alternative methods for reestimating the between cluster and residual variances in a blinded manner, and why we believe our chosen approach should be preferred.

Simulation study

With the above considerations, a SSRE trial design scenario is fully specified given , where Here, is the vector of period effects, and B is a binary indicator variable that takes the value 1 if blinded SSRE is utilized, and the value 0 if unblinded SSRE is utilized. Finally, is the indicator function on event A. Given we can simulate a SW‐CRT utilizing this SSRE procedure by generating random multivariate normal observations. From this, the empirical rejection rate (ERR) of a particular scenario can be estimated by performing a large number of replicates simulations. To this end, define to be 1 if the result of replicate s of a trial simulated according to scenario is to reject H 0, and 0 otherwise. For any number of replicates r, the ERR for scenario is Similarly, we record the values of , the total sample size required in replicate s, computed using and , the initial and reestimated per cluster per period sample sizes for this replicate. This allows us to examine the distribution of the total required sample size, which we denote from here by . In this article, for all considered scenarios. In what follows, we consider performance in a wide variety of scenarios. However, many of the parameters in remain fixed. In particular, they are set based on two motivating trial design scenarios. Firstly, Bashour et al. (2013) conducted a SW‐CRT to assess the effect of training doctors in communication skills on women's satisfaction with doctor‐woman relationship during labour and delivery. The trial utilized a balanced complete block SW‐CRT design, enrolling four hospitals, and gathering data over five time periods. The final analysis estimated the between cluster and residual variances to be and , respectively. For these variance parameters, the utilised design would have required 70 patients per cluster per time period for the trials desired type‐I and type‐II error rates of 0.05 and 0.1, respectively, when powering for a clinically relevant difference of 0.2, using the methods above. Thus in Trial Design Setting (TDS) 1 we fix , , , , and . Moreover, , , and X is such that a single cluster switches to the experimental intervention in time periods two through five. The parameters of TDS2 are based upon the typical characteristics of SW‐CRTs according to a recent review (Grayling et al., 2017a). Precisely, adapting Grayling et al. (2017b), we set and , in order to consider a more modest value for the ICC of . In addition, we set , , , , , and X such that three clusters switch to the experimental intervention in time periods two through five, and two clusters in time periods six through nine. This implies that the actual total sample size required by this trial is approximately that accrued on average in SW‐CRTs completed to date. For simplicity, in both TDSs we take . We therefore consider the effect of different choices for t, , , , , τ, , and B. We in general assume that . That is, each variance component is either underspecified by 50%, correctly specified, or overspecified by 50%. Moreover, we consider three possible combinations of values for , When and 1,000, so there is no practical limit on . When and 1,000, so there is no practical upper limit on , but it must be at least as large as the initially specified per cluster per period sample size. When and , so there is no lower limit on , but it cannot be larger than the initially specified per cluster per period sample size. The upper limits of 1,000 are retained simply for computational reasons, as memory allocation issues can occur when is extremely large. Note that given in all instances, the trial will never be terminated at the reestimation point, with all clusters recruiting at least one participant in the subsequent time periods. Moreover, results are presented here only for the first scenario with and 1,000. Our findings for the other two scenarios are provided in the Supplementary Material. Software to perform our simulations is available from https://github.com/mjg211/article_code.

RESULTS

Performance for varying and

To begin, we consider how the SSRE procedures perform as and are misspecified to varying degrees. Precisely, we set and for TDS1 and TDS2, respectively, and explore , with ; giving the empirical type‐I error rate (ETI), or ; giving the empirical power (EP). For the blinded procedure, we take , and here we consider only the case where and 1,000. Our findings are presented in Table 1, which displays the ERRs of the reestimation procedures and the corresponding fixed sample SW‐CRT design. Furthermore, in Table 2 the median value of in each case is listed. Additionally, Supporting Information Figures 1, 2, 4, and 5 together depict the distributions of and when employing blinded or unblinded reestimation, when and . Similarly, Supporting Information Figures 3 and 6 display the corresponding distributions of .
Table 1

Empirical type‐I error rates () and power () of the blinded () and unblinded reestimation procedures, along with the corresponding fixed sample SW‐CRT design are shown

τ=0 τ=δ
σc2 σe2 BlindedUnblindedFixedBlindedUnblindedFixed
Trial Design Setting 1
0.5σc2 0.5σe2 0.05810.05760.06460.88120.87990.6922
0.5σc2 σe2 0.05890.06010.05850.88590.88580.8929
0.5σc2 1.5σe2 0.06320.06410.05690.89350.89630.9700
σc2 0.5σe2 0.05680.05630.06270.88170.87880.7051
σc2 σe2 0.05930.06190.06000.88480.88430.9034
σc2 1.5σe2 0.06430.06300.05670.89570.89680.9736
1.5σc2 0.5σe2 0.05830.05640.06210.88250.88040.7084
1.5σc2 σe2 0.05940.05870.05940.88620.88640.9067
1.5σc2 1.5σe2 0.06300.06420.05670.89740.89540.9736
Trial Design Setting 2
0.5σc2 0.5σe2 0.02540.02700.02660.82820.80020.5875
0.5σc2 σe2 0.02600.02610.02550.82840.80680.8024
0.5σc2 1.5σe2 0.02690.02610.02590.82840.80940.9125
σc2 0.5σe2 0.02600.02630.02710.82950.80060.5910
σc2 σe2 0.02710.02740.02570.82830.80590.8021
σc2 1.5σe2 0.02620.02740.02530.83010.81230.9333
1.5σc2 0.5σe2 0.02660.02730.02580.82610.80260.5903
1.5σc2 σe2 0.02580.02580.02550.82880.80530.8486
1.5σc2 1.5σe2 0.02690.02680.02540.82870.80950.9348

Results are given for Trial Design Settings 1 () and 2 (), for a selection of possible values for the assumed variance parameters, with and 1,000.

Table 2

Median values of the final total required sample size () are shown for and using the blinded () and unblinded reestimation procedures, along with the corresponding fixed sample SW‐CRT design

τ=0 τ=δ
σc2 σe2 BlindedUnblindedFixedBlindedUnblindedFixed
Trial Design Setting 1
0.5σc2 0.5σe2 1,5561,5567001,6121,556700
0.5σc2 σe2 1,3641,3561,3401,3961,3561,340
0.5σc2 1.5σe2 1,3761,3762,0401,4081,3762,040
σc2 0.5σe2 1,5521,5447201,6001,544720
σc2 σe2 1,3521,3521,4001,3921,3521,400
σc2 1.5σe2 1,3841,3842,1201,4161,3842,120
1.5σc2 0.5σe2 1,5401,5327401,5961,532740
1.5σc2 σe2 1,3481,3481,4201,3881,3481,420
1.5σc2 1.5σe2 1,3881,3882,1401,4201,3882,140
Trial Design Setting 2
0.5σc2 0.5σe2 1,4401,4407201,6001,440720
0.5σc2 σe2 1,2601,2601,2601,3401,2601,260
0.5σc2 1.5σe2 1,2401,2401,8001,3201,2401,800
σc2 0.5σe2 1,4401,4407201,6001,440720
σc2 σe2 1,2601,2601,2601,3401,2601,260
σc2 1.5σe2 1,2601,2601,9801,3401,2601,980
1.5σc2 0.5σe2 1,4401,4407201,6001,440720
1.5σc2 σe2 1,2801,2801,4401,3601,2801,440
1.5σc2 1.5σe2 1,2601,2601,9801,3401,2601,980

Results are given for Trial Design Settings 1 () and 2 (), for a selection of possible values for the assumed variance parameters, with and 1,000.

Empirical type‐I error rates () and power () of the blinded () and unblinded reestimation procedures, along with the corresponding fixed sample SW‐CRT design are shown Results are given for Trial Design Settings 1 () and 2 (), for a selection of possible values for the assumed variance parameters, with and 1,000. Median values of the final total required sample size () are shown for and using the blinded () and unblinded reestimation procedures, along with the corresponding fixed sample SW‐CRT design Results are given for Trial Design Settings 1 () and 2 (), for a selection of possible values for the assumed variance parameters, with and 1,000. In general, for the fixed design, assuming larger values for the variance parameters leads to an increased EP, a decreased ETI, and larger requisite final sample sizes, as would be expected. While this appears to be true for the EP in the SSRE designs, it is not always the case for the ETI. However, assuming larger values for the variance parameters does lead to improved performance of the SSRE procedures in terms of estimating and at the interim analysis. This is true both in terms of the median reestimated values, and the observed variability in the estimates (Supporting Information Figures 1, 2, 4, and 5). It should be noted that, as we would expect, the blinded reestimation procedure tends to perform worse at estimating when than (Supporting Information Figures 1, 2, 4, and 5), as the blinded estimator is only unbiased when . Moreover, for the unblinded procedure tends to underestimate the value of , while the blinded procedure overestimates its value. In TDS1, for certain values of the assumed variance parameters there is large inflation of the ETI above the nominal level. This is an issue common to both the SSRE procedures and the fixed design, with the maximal inflation observed for , in the unblinded reestimation design, where the ETI is 0.0653. Finally, for TDS1, the blinded and unblinded procedures have similar values for the EP. However, from Table 2 we can say that the blinded procedure routinely requires a larger number of measurements than its unblinded analog. In contrast, for TDS2 there is only small inflation to the ETI, with the blinded and unblinded methods displaying similar ETIs. In this case, the blinded procedure always has a larger EP than the unblinded procedure. Examining Table 2, it is clear why this is the case, as the blinded procedure has a higher median value for , which is in turn a consequence of our observation above on the blinded procedures propensity to overestimate . Given that the unblinded procedure always attains the desired power, this is arguably a disadvantage of the blinded approach. Overall, it is clear that when the value of is underspecified, the SSRE procedures generally have a far higher EP than the corresponding fixed SW‐CRT design, with comparable if not preferable ETIs. For example, when and in TDS2, the blinded procedure has an EP of 0.8282, while the corresponding conventional SW‐CRT design has an EP of only 0.5875; an increase of 41%. The reason for this is clear from Table 2, as we can see the blinded procedure is able to effectively utilize the interim estimates of the variance parameters to increase the final requisite sample size. Similarly, when the variance parameters are overspecified, the SSRE procedures are able to reduce the total requisite sample size by taking , and bring the EP closer to the desired level.

Performance for varying t

Next, we assess the impact upon the ETI and EP of the choice of the SSRE point t. As above, we set for the blinded procedure, and take and 1,000. However, we now focus only on the cases where . Table 3 displays the ETI and EP of the blinded and unblinded SSRE procedures when and , respectively, for in TDS1 and in TDS2. In addition, Table 4 displays the corresponding median values of . Finally, Supporting Information Figures 7– 12 display the distributions of , , and , for these scenarios.
Table 3

Empirical type‐I error rates () and power () of the blinded () and unblinded reestimation procedures are shown

τ=0 τ=δ
Trial Design Setting 1
Procedure σc2 σe2 t=2 t=3 t=4 t=2 t=3 t=4
Blinded 0.5σc2 0.5σe2 0.05920.05810.05120.87020.88120.8151
Blinded σc2 σe2 0.05970.05930.06170.88510.88480.8971
Blinded 1.5σc2 1.5σe2 0.05700.06300.06180.88780.89740.9541
Unblinded 0.5σc2 0.5σe2 0.06290.06040.05030.86930.87990.8165
Unblinded σc2 σe2 0.05990.06520.06430.88400.88430.8966
Unblinded 1.5σc2 1.5σe2 0.05940.06410.06240.88660.89550.9551
Trial Design Setting 2
Procedure σc2 σe2 t=3 t=5 t=7 t=3 t=5 t=7
Blinded 0.5σc2 0.5σe2 0.02630.02540.02490.81400.82820.8146
Blinded σc2 σe2 0.02610.02710.02660.81590.82830.8271
Blinded 1.5σc2 1.5σe2 0.02540.02690.02620.81810.82870.8883
Unblinded 0.5σc2 0.5σe2 0.02680.02700.02580.80380.80020.8003
Unblinded σc2 σe2 0.02740.02680.02670.81300.80590.8030
Unblinded 1.5σc2 1.5σe2 0.02560.02620.02620.81270.80950.8891

Results are given for Trial Design Settings 1 and 2, for a selection of possible values for the assumed variance parameters, and as a function of the reestimation time point t, when and 1,000.

Table 4

Median values of the final total required sample size () required by the blinded () and unblinded reestimation procedures are shown

τ=0 τ=δ
Trial Design Setting 1
Procedure σc2 σe2 t=2 t=3 t=4 t=2 t=3 t=4
Blinded 0.5σc2 0.5σe2 1,3121,5564,5601,3481,6124,560
Blinded σc2 σe2 1,3521,3521,3361,3761,3921,392
Blinded 1.5σc2 1.5σe2 1,4561,3881,7161,4681,4201,716
Unblinded 0.5σc2 0.5σe2 1,3001,5564,5601,3001,5564,560
Unblinded σc2 σe2 1,3521,3521,3281,3521,3521,332
Unblinded 1.5σc2 1.5σe2 1,4561,3881,7161,4561,3881,716
Trial Design Setting 2
Procedure σc2 σe2 t=3 t=5 t=7 t=3 t=5 t=7
Blinded 0.5σc2 0.5σe2 1,3201,4401,9601,3201,6002,160
Blinded σc2 σe2 1,2601,2601,2601,3801,3401,380
Blinded 1.5σc2 1.5σe2 1,3801,2601,5801,3801,3401,580
Unblinded 0.5σc2 0.5σe2 1,3201,4402,0001,3201,4402,000
Unblinded σc2 σe2 1,2601,2601,2601,2601,2601,260
Unblinded 1.5σc2 1.5σe2 1,3801,2601,5801,3801,2601,580

Results are given for Trial Design Settings 1 and 2, for a selection of possible values for the assumed variance parameters, and as a function of the reestimation time point t, when and 1,000.

Empirical type‐I error rates () and power () of the blinded () and unblinded reestimation procedures are shown Results are given for Trial Design Settings 1 and 2, for a selection of possible values for the assumed variance parameters, and as a function of the reestimation time point t, when and 1,000. Median values of the final total required sample size () required by the blinded () and unblinded reestimation procedures are shown Results are given for Trial Design Settings 1 and 2, for a selection of possible values for the assumed variance parameters, and as a function of the reestimation time point t, when and 1,000. We observe no clear trend to the ETI as t is increased in either TDS. In TDS2, the ETI is comparable for each t, for each of the different assumed combinations of variance parameters. However, in TDS1, the ETI is more sensitive to the choice of t. As in Section 3.1, this is likely a consequence of the smaller number of clusters for this TDS. In TDS2, each of the values of t confers the largest EP for some combination of the variance parameters. While in TDS1, or provide the largest power in at least one considered case. Nonetheless, in some instances, it is clear that placing the reestimation point later in to the trial can cause a substantial loss of power. For example, when and in TDS1, the unblinded procedure has an EP of 0.8799 when , but this drops to 0.8165 for . In addition, Table 4 and Supporting Information Figures 9 and 12 provide a clear warning as to the possible issues with utilizing larger values of t: when the variance parameters are underspecified this can lead to far larger sample sizes being required, as extremely large values of are required to attempt to attain the desired power. Finally, examining Supporting Information Figures 7, 8, 11, and 12, we can see that, as would be anticipated, the reestimation procedures are generally able to more accurately estimate the variance components as t is increased. The exception to this rule is the blinded procedure when . In this case, larger values of t increase the median bias in .

Performance for varying

In this section, we consider the effect the value of has on the blinded re‐estimation procedure. We set and for TDS1 and TDS2, respectively, and explore , with and , when and 1,000. Two values of are considered: and . Our results are presented in Table 5, which contains the ERRs, and in Supporting Information Figures 13– 16, which display the distributions of and .
Table 5

Empirical type‐I error rates () and power () of the blinded reestimation procedures are shown

τ=0 τ=δ
σc2 σe2 τ=0 τ=δ τ=0 τ=δ
Trial Design Setting 1
0.5σc2 0.5σe2 0.05810.05810.88120.8739
0.5σc2 σe2 0.05890.05960.88590.8779
0.5σc2 1.5σe2 0.06320.06340.89350.8905
σc2 0.5σe2 0.05680.05730.88170.8739
σc2 σe2 0.05930.06090.88480.8798
σc2 1.5σe2 0.06430.06260.89570.8961
1.5σc2 0.5σe2 0.05830.05630.88250.8736
1.5σc2 σe2 0.05940.05840.88620.8801
1.5σc2 1.5σe2 0.06300.06420.89740.8944
Trial Design Setting 2
0.5σc2 0.5σe2 0.02540.02620.82820.7876
0.5σc2 σe2 0.02600.02540.82840.7996
0.5σc2 1.5σe2 0.02690.02600.82840.8026
σc2 0.5σe2 0.02600.02630.82950.7880
σc2 σe2 0.02710.02630.82830.7995
σc2 1.5σe2 0.02620.02690.83010.8096
1.5σc2 0.5σe2 0.02660.02580.82610.7881
1.5σc2 σe2 0.02580.02510.82890.7991
1.5σc2 1.5σe2 0.02690.02690.82870.8071

Results are given for Trial Design Settings 1 () and 2 (), for a selection of possible values for the assumed variance parameters, and as a function of , when and 1,000.

Empirical type‐I error rates () and power () of the blinded reestimation procedures are shown Results are given for Trial Design Settings 1 () and 2 (), for a selection of possible values for the assumed variance parameters, and as a function of , when and 1,000. We can see that the value of appears to have little influence on the ETI for either TDS. However, in some cases the choice of has a notable effect on the EP, with specifically reducing the EP. This is especially evident in TDS2. From Supporting Information Figures 13–16 we can see why this is the case. When , on average, reduced values for are obtained compared to choosing . Thus results in larger final requisite sample sizes, which provides an increase to the EP.

DISCUSSION

In this article, we have presented blinded and unblinded SSRE procedures for cross‐sectional SW‐CRTs. These methods should assist with scenarios in which there is difficultly in determining a trial's required sample size because of the need to specify values for several nuisance parameters. We were able to demonstrate that, at least for the considered scenarios, the SSRE procedures could increase power substantially over a conventional SW‐CRT design when the variance parameters were underspecified. Unfortunately, in TDS1 there were instances of substantial inflation to the ETI rate using our SSRE procedures (Table 1). This was not surprising given the extremely low number of clusters in this scenario, with past research highlighting issues in such settings (Taljaard, Teerenstra, Ivers, & Fergusson, 2016; Grayling et al., 2017b). Additionally, it follows results observed for parallel‐group CRTs (van Schie & Moerbeek, 2014). To address this, one could look to use the Kenward‐Roger correction for performing a hypothesis test on a fixed effect in a linear‐mixed model (Kenward & Roger, 1997). In practice this would be expected to be useful, as it was indeed recently demonstrated to be for parallel group CRTs (Kahan et al., 2016). However, it is extremely computationally intensive to examine the utility of this within the context of a large scale SSRE simulation study. We thus omit such considerations here. Alternatively, an alpha‐level adjustment procedure, as considered for example by Golkowski et al. (2014) and Friede and Stammer (2010), among others, could assist in controlling the type‐I error rate more accurately. This approach seeks to identify the largest possible inflation that could occur, and lowers the value of α used in the final test accordingly to control to the nominal level for this worst case scenario. Of course, when utilizing either of these approaches one must bare in mind that a penalty to the EP would likely be observed. Moreover, in TDS1 the SSRE procedures routinely did not display an EP above the nominal level (Table 1). To combat this, one could employ a sample size inflation factor, as proposed by Zucker, Wittes, Schabenberger, and Brittan (1999). This has been demonstrated to be highly effective in a range of trial design settings (e.g., Friede & Kieser, 2013; Golkowski et al., 2014). Nonetheless, it was clear that though the SSRE procedure did not in many instances precisely meet the desired operating characteristics, their performance in comparison to the fixed sample design was often impressive. When the variance parameters were overspecified, setting allowed the SSRE procedures to reduce the group size and attain a power closer to the nominal level. Similarly statements hold in regard to when the variance parameters were underspecified having set . Thus the SSRE designs may have real utility in combating issues of uncertainty over and at the design stage, especially if an adjustment procedure (like those described above) is incorporated to enhance their ability to attain the desired type‐I and type‐II error rates. Choosing between the blinded and unblinded procedures may be more challenging however. As seen in Table 2, the blinded procedure typically requires a larger sample size for only a small gain in power over an unblinded option. This, in particular, would need to be taken into account to choose the preferred reestimation approach on statistical grounds. Of course, in practice, there are further important operational details that must be taken into account in order to make a choice between the blinded and unblinded procedures. For example, utilizing unblinded reestimation can induce additional complexities. In particular, an independent data monitoring committee may then be required, in order to ensure that study personnel remain blinded. In my opinion, it is important to highlight that this is not only a choice between statistical operating characteristics, but also a decision for which operational aspects have to be taken into account. An unblinded sample size reestimation needs careful planning, and potentially an independent data monitoring committee, to assure that the study personnel stays blinded. From Table 1 and Supporting Information Table 1 in particular, it is clear that altering the value of , for fixed , has little effect on the performance of the reestimation procedures. As noted by one of the anonymous reviewers, it is possible, however, that this may not be the case when . We explore such scenarios in the Supplementary Material. Ultimately, we find that the performance of the reestimation procedures is insensitive to the values of and even when . In addition, given the fact that will likely in general be biased for when using the blinded procedure, it is reasonable to ask how well the blinded procedure could perform if we simply assume that . We also consider such scenarios in the Supplementary Material. From this, we caution against assuming the between cluster variance is not of importance, as the trial operating characteristics can be severely effected if this is not the case. We observed that the ETI and EP were similar for several choices of t, particularly in TDS1, but the EP was sometimes substantially lower if the reestimation point was late in the trial (Table 3). This should not be surprising. There is clearly a trade‐off to be made in terms of how late in the trial one places the reestimation point. The longer we wait, the more accurately we are in general able to estimate the variance parameters. However, we then have less time to readjust for any pretrial misspecification. Indeed, this point is a complex one, related to recent research on the value of each cluster‐period in a SW‐CRT (Kasza & Forbes, 2017), and how horizontal and vertical components contribute to the treatment effect estimate in a SW‐CRT (Matthews & Forbes, 2017). As an example, consider the choice in TDS1. This leaves a single time period left in which to carry out the trial with a modified sample size. However, with our choice of X, in this final time period all of the clusters receive the experimental intervention. The implication of this is that the majority of the cluster‐periods in time period five actually contribute little to estimating the treatment effect. Consequently, even taking 1,000 may not allow the desired power to be attained if the preceding time periods were conducted with too small a sample size. Accordingly, one may therefore suggest an intermediate value for t, such as in TDS1, to be preferable in most cases. From Section S.M.7 in the Supplementary Material, it is clear that when specifying a SSRE procedure, it is important to choose the values of and carefully. In particular, while it may be preferable to have , this could have negative consequences upon the EP. However, this does confer an advantage that when the SSRE procedures were able to reduce the power to closer to the nominal level. Likewise, increasing the value of may seem beneficial, but one then needs both to be able to find more patients to recruit in the later periods, and also to be able to logistically handle a larger sample size. This may be a problem particularly for scenarios where the SW‐CRT design is being utilized because of resource constraints. Similarly, if using the blinded SSRE procedure, the value for should be chosen with care. From Table 5, it was evident that this can have consequences on, in particular, the EP. As has been previously noted for conventional parallel arm trials (Kieser & Friede, 2003), the bias introduced into the estimate of when , may make the choice preferable. However, one should be careful not to automatically assume this is the case. Specifically, the choice will likely often provide greater power than the choice . But, in certain cases may in its own right provide the desired power, as is approximately the case for TDS2 by Table 5. In this instance, the increased power attained by choosing could be disadvantageous because of the associated increased required sample sizes. As general guidance, one may anticipate that would be preferable for settings with small numbers of clusters, where attaining the desired power may be more challenging, as seen for TDS1. An additional point of note is our method for specifying in the blinded re‐estimation procedure. Explicitly, we choose to set when . It may appear more logical to set in this instance. We made our choice based on the fact that it would be rare for one to truly believe that . Consequently, we felt would more likely indicate that an overly large value for had been chosen. In practice though, this specification would likely have little influence on the performance of the blinded reestimation procedure unless the magnitude of is large. Note also that throughout this work we considered only cases with . However, the SSRE procedures considered here are only asymptotically invariant to the value of the period effects. It would thus in general be important to assess the effect of nonzero period effects on the SSRE procedures, though it is reasonable to anticipate that it will be small. There are several practical factors that must be considered before SSRE is incorporated into a SW‐CRT design. Primarily, our methodology is dependent upon data from all clusters being available for analysis immediately following period t. The efficiency of the procedures would suffer if this were not the case. Therefore, it would be important for measures to be put in place for efficient data collection, storage, and analysis. In addition, there may be some instances where SSRE is not realistic. For example, if the intervention was a planned roll‐out that is part of a larger programme implementation. A trialist must consider their scenario carefully before utilizing SSRE. Several possible extensions to our procedures are possible. Firstly, we here only addressed cross‐sectional SW‐CRT designs analyzed with the Hussey and Hughes model. Though the majority of SW‐CRT research has been set in this domain, it would be beneficial to also establish methods to incorporate SSRE in to cohort designed SW‐CRT, different endpoints of interest, or indeed different analysis models. While it would be relatively simple to explore the performance of an unblinded procedure in these settings, methodology for blinded reestimation would be more complex. Similar statements also hold for allowing variable cluster sizes, and also incorporating the interim estimated value for τ in to the re‐estimation procedure. Additionally, we considered here a scenario in which the number of clusters remained fixed throughout the trial; adjusting only the per cluster per period sample size following the reestimation point. One could also explore the performance of a procedure that increases the value of C following reestimation, creating an incomplete‐block SW‐CRT. For scenarios in which patients are hard to come by, but clusters are not, this would be a useful extension. It is worth noting that our procedures are actually applicable to any cross‐sectional CRT design to be analyzed with the Hussey and Hughes model. This means, for example, that it would allow also the incorporation of SSRE in to a cluster randomised crossover trial, which is being increasingly acknowledged in the trials community as a useful design (Arnup et al., 2014). Regardless of the practical considerations discussed above, and the possible future avenues of extension to our methods, it is clear that the ability to include a SSRE point in to SW‐CRT designs is a useful addition to the methodologists toolbox.

CONFLICT OF INTEREST

The authors have declared no conflict of interest. Supplementary Material Click here for additional data file. Supplementary Material Click here for additional data file.
  25 in total

1.  Blinded sample size recalculation in multicentre trials with normally distributed outcome.

Authors:  Katrin Jensen; Meinhard Kieser
Journal:  Biom J       Date:  2010-06       Impact factor: 2.207

2.  Interim analyses for monitoring clinical trials that do not materially affect the type I error rate.

Authors:  A L Gould
Journal:  Stat Med       Date:  1992-01-15       Impact factor: 2.373

Review 3.  Sample size recalculation in internal pilot study designs: a review.

Authors:  Tim Friede; Meinhard Kieser
Journal:  Biom J       Date:  2006-08       Impact factor: 2.207

4.  Stepped wedge designs: insights from a design of experiments perspective.

Authors:  J N S Matthews; A B Forbes
Journal:  Stat Med       Date:  2017-08-08       Impact factor: 2.373

5.  The role of internal pilot studies in increasing the efficiency of clinical trials.

Authors:  J Wittes; E Brittain
Journal:  Stat Med       Date:  1990 Jan-Feb       Impact factor: 2.373

6.  Sample size calculation for stepped wedge and other longitudinal cluster randomised trials.

Authors:  Richard Hooper; Steven Teerenstra; Esther de Hoop; Sandra Eldridge
Journal:  Stat Med       Date:  2016-06-28       Impact factor: 2.373

7.  Substantial risks associated with few clusters in cluster randomized and stepped wedge designs.

Authors:  Monica Taljaard; Steven Teerenstra; Noah M Ivers; Dean A Fergusson
Journal:  Clin Trials       Date:  2016-03-03       Impact factor: 2.486

Review 8.  The stepped wedge trial design: a systematic review.

Authors:  Celia A Brown; Richard J Lilford
Journal:  BMC Med Res Methodol       Date:  2006-11-08       Impact factor: 4.615

9.  The effect of training doctors in communication skills on women's satisfaction with doctor-woman relationship during labour and delivery: a stepped wedge cluster randomised trial in Damascus.

Authors:  Hyam N Bashour; Mona Kanaan; Mayada H Kharouf; Asmaa A Abdulsalam; Mohammed A Tabbaa; Salah A Cheikha
Journal:  BMJ Open       Date:  2013-08-14       Impact factor: 2.692

10.  Blinded and unblinded sample size reestimation procedures for stepped-wedge cluster randomized trials.

Authors:  Michael J Grayling; Adrian P Mander; James M S Wason
Journal:  Biom J       Date:  2018-08-03       Impact factor: 2.207

View more
  4 in total

1.  Sample size and power considerations for cluster randomized trials with count outcomes subject to right truncation.

Authors:  Fan Li; Guangyu Tong
Journal:  Biom J       Date:  2021-03-10       Impact factor: 1.715

2.  Blinded and unblinded sample size reestimation procedures for stepped-wedge cluster randomized trials.

Authors:  Michael J Grayling; Adrian P Mander; James M S Wason
Journal:  Biom J       Date:  2018-08-03       Impact factor: 2.207

3.  Response adaptive intervention allocation in stepped-wedge cluster randomized trials.

Authors:  Michael J Grayling; James M S Wason; Sofía S Villar
Journal:  Stat Med       Date:  2022-01-21       Impact factor: 2.497

4.  Interim data monitoring in cluster randomised trials: Practical issues and a case study.

Authors:  K Hemming; J Martin; I Gallos; A Coomarasamy; L Middleton
Journal:  Clin Trials       Date:  2021-06-22       Impact factor: 2.486

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.