Literature DB >> 31549566

Testing multiple dose combinations in clinical trials.

Saswati Saha1, Werner Brannath1, Björn Bornkamp2.   

Abstract

Drug combination trials are often motivated by the fact that individual drugs target the same disease but via different routes. A combination of such drugs may then have an overall better effect than the individual treatments which has to be verified by clinical trials. Several statistical methods have been explored that discuss the problem of comparing a fixed-dose combination therapy to each of its components. But an extension of these approaches to multiple dose combinations can be difficult and is not yet fully investigated. In this paper, we propose two approaches by which one can provide confirmatory assurance with familywise error rate control, that the combination of two drugs at differing doses is more effective than either component doses alone. These approaches involve multiple comparisons in multilevel factorial designs where the type 1 error can be controlled first, by bootstrapping tests, and second, by considering the least favorable null configurations for a family of union intersection tests. The main advantage of the new approaches is that their implementation is simple. The implementation of these new approaches is illustrated with a real data example from a blood pressure reduction trial. Extensive simulations are also conducted to evaluate the new approaches and benchmark them with existing ones. We also present an illustration of the relationship between the different approaches. We observed that the bootstrap provided some power advantages over the other approaches with the disadvantage that there may be some error rate inflation for small sample sizes.

Entities:  

Keywords:  Drug combination; factorial design; intersection union tests; multiple testing

Mesh:

Year:  2019        PMID: 31549566      PMCID: PMC7309363          DOI: 10.1177/0962280219871969

Source DB:  PubMed          Journal:  Stat Methods Med Res        ISSN: 0962-2802            Impact factor:   3.021


1 Introduction

Combining different drugs is an important treatment option in many therapeutic areas such as respiratory, cardiovascular disease, cancer or infectious diseases. The hope is that by combining two drugs (with typically different modes of action), one can achieve a greater beneficial effect than either therapy alone. According to the regulatory requirement by the U.S. Food and Drug Administration's policy (21 CFR 300.50,[1] CDER, 20132), the fixed-dose combination drug must have confirmatory evidence for being more effective than each component drug alone. The primary questions that often arise in a drug combination therapy trial are: (1) Does there exist a dose combination that shows a better effect than the placebo control (effectiveness)? (2) Does there exist a dose combination that is superior to the individual treatments (superiority), where the individual treatments are often termed as monotherapies? (3) What are the specific combinations that fulfill both effectiveness and superiority? Laska and Meisner[3] and Snapinn[4] are the first to consider the problem of testing the superiority of a certain combination treatment over the component treatments in a single dose combination setting. They conducted a “min-test”, where the minimum of the test statistics comparing the combination treatment with the monotherapies are used to show that the combination treatment is better. Extending the above approach in a multiple dose combination setting is not simple. In a multiple dose combination trial, multi-level factorial designs involving simultaneous multiple dose combination comparison comes into play and one has to propose multiple testing procedure (MTP) to address the multiplicity issues here. Several authors have addressed the primary questions stated in the earlier paragraph in several ways. Hung[5-7] proposed two single-step testing procedures that showed that there exists at least one combination in a multiple dose factorial drug combination study that is better than administering the component drugs alone. Hellmich and Lehmacher[8] proposed a testing method for identifying the set of all minimum effective combinations in the case of monotone mean responses. Buchheister and Lehmacher[9] proposed a closed testing procedure using special linear contrast tests and extended the global maximum test by Hung[6] to a local maximum test for the identification of the superior dose combinations while preserving the family-wise error rate. Soulakova and Sampson[10] and Soulakova[11] proposed a procedure where their objective was to identify the set of minimum effective combination doses using the global average tests proposed by Hung[5] under a closed testing principle. In all of these papers, the problems of efficacy and superiority of a combination drug were addressed separately which allowed explicit discussion of statistical issues. However, in practice, if a certain effective combination is shown to be superior, it is necessary to explain how results of the two individual procedures may be combined and what adjustments are needed to claim both efficacy and superiority. To address this, Soulakova[12,13] expressed the problem of identifying the effective and superior drug combinations as a two stage problem, where the min-test is conducted at the first stage for comparing the drug combinations with the monotherapies at each dose level and then Holm's rejective multiple testing approach[14] is employed in the second stage to obtain doses that show superiority over placebo. In this paper, we focus on showing superiority with regard to the monotherapies and ignore the formal requirement of showing effectiveness, i.e. superiority over placebo. However, we will indicate in the discussion how to include the tests for effectiveness in some of the approaches considered here. We have already seen that many approaches are suggested in the literature to address the goals of testing superiority of drug combinations in a multi-level factorial design and identifying the set of superior dose combinations. The method suggested by Hung[7] is one of the pioneer approaches that proposed a global test to deal with the above problems. Other authors have mainly proposed alternatives to test the same global null hypothesis that there exists at least one combination that provides superiority. Some authors proposed alternative approaches[9,15] to control the family-wise error rate (FWER) strongly and identify all beneficial combinations in a multiple dose combination setting. However, most of the approaches proposed are either based on step-wise MTPs for a nonhierarchical hypothesis family, such as methods by Holm,[14] Hochberg,[16] and resampling methods by Westfall and Young[17]; see Soulakova[12,13] or rely on closed testing principles proposed by Marcus, Peritz, and Gabriel[18]; also see Hellmich and Lehmacher,[8] Buchheister and Lehmacher,[9] Soulakova and Sampson,[10] and Soulakova.[11] In this article, instead of relying on conventional MTPs mentioned earlier (like Holm[14] or Hochberg[16]), we will propose two new multiple testing procedures, by which one can test for superiority of the drug combination using: (i) a parametric bootstrap approach and (ii) FWER control under the least favorable null configuration. The parametric bootstrap approach suggested here estimates the parameters from the given dataset under the constraints imposed by the null hypothesis and obtains the null distribution of the test statistics by sampling data with the estimated parameters. Hence, it provides a tool to carry out the multiple comparisons in a parametric setup, without worrying about all the sampling distributions of the inherent test statistics in the composite null hypotheses. The least favorable null configuration approach aims to control the maximum type 1 error rate in the above multiple testing problem. It identifies the worst possible configurations (a subset of the null parameter space) that allow one to obtain a bound on the size of the test and control it within the desired significance limit. Thus the bootstrap approach and the least favorable null configuration approach both suggest ways of controlling the FWER which is a mandatory requirement for therapeutic dose response studies in Phase III clinical trials. While the least favorable null configuration approach leads to a FWER that is always below the nominal level α (and often much smaller), the bootstrap approach controls the FWER only asymptotically. However, the latter has the advantage of being less conservative.

2 Problem

Consider a random vector Y, containing the clinical measurement of interest and a (r + 1) × (s + 1) factorial design trial where the dose levels are coded as 0, 1, 2,…r for drug A and 0, 1, 2,…s for drug B. The response Y is observed for (r + 1) × (s + 1) parallel dose combination groups and it is assumed to have the following model where k = 1, 2,…n, i = 0, 1,…r and j = 0, 1,…s. Here {μ, i = 1, 2,…r} are the mean responses of the monotherapies of Drug A and {μ0, j = 1, 2,…s} are the mean responses of the monotherapies of Drug B and {μ, i = 1, 2,…r, j = 1, 2,…s} denote the mean response at dose combination, dose i of drug A and dose j of drug B. We assume here that . For the (i, j)th dose combination, the alternative hypothesis of interest is that the dose combination is more effective than both monotherapies, i.e. H1: μ > μ and μ > μ0. The corresponding null hypothesis is H0: μ ≤ μ or μ ≤ μ0. The global hypothesis associated with testing all active dose combinations versus their respective components is

3 Methods

3.1 Max-min test

We can rewrite the global null and corresponding alternative of the hypotheses discussed in equation (2) as If T denotes the test statistic for testing H0 against H1, then the test statistic for testing the global null is given by where and are the contrast test statistics used for testing whether drug combination is superior to the monotherapies with respect to Drug A and Drug B, respectively. As suggested by Hung[7] and Soulakova,[12] a simple approach here is to compute the p-value for each T using the minimum of the test statistics and . The distribution of is given below where and are the non-central t distributions with non-centrality parameter and , respectively, and with degrees of freedom . The raw p-values (p) for testing H0 can be easily obtained but for testing multiple combinations; these raw p-values need to be adjusted. Multiplicity adjustment is challenging because the null distribution of the “Max” test statistic in equation (4) is unknown and it is difficult to compute. Essentially one needs to adjust the p such that the FWER is controlled at significance level α for any of the possible null configurations. We can compute the p-values for all H0 and perform a Bonferroni correction for the (r × s) union tests shown in equation (3). However, the Bonferroni method is over conservative in almost all situations, so we consider alternative approaches to control the type 1 error rate. A more efficient approach is to control the maximum type 1 error based on the joint distribution of the test statistics. The maximum type 1 error can be calculated by searching for the “worst possible parameter configurations” within the null space for which the size of the test is maximized.[19] We will see below that in our case the maximum type I error is not achieved by any finite parameter configurations but for hypothetical limiting cases where each T becomes equal to either or because the other test statistics becomes infinitely large. This is further elaborated in the next paragraph. These limiting configurations are denoted by least favorable configurations (LFC) in the rest of the article. Since the LFC are impossible in reality, we suggest an alternative approach, where the mean parameters () are estimated under the null constraint (3) and these null space restricted estimates are utilized via bootstrapping to obtain the critical value for the above multiple testing problem. This method gives more realistic estimates of the type 1 error rate. Note that the above multiple testing procedure is based on test statistics that satisfy the subset pivotality condition.[17] The subset pivotality conditions asserts; if , where G denotes the set of all active dose combinations and K is a subset of G where H0 is true for all (i, j) ∈ K, then the test statistics T, for some (i, j) ∈ K depend on the nuisance parameter and sample sizes n, n and n0 but not on the remaining parameters . The distribution of is therefore the same under the complete hypothesis and the reduced hypothesis . Furthermore, each hypothesis is tested with a “Max” statistics and according to Westfall and Troendle[20] both approaches, LFC and bootstrap approach, attain strong control of FWER. The following section elaborates how the multiplicity issue is dealt in the above testing problem.

3.1.1 Least favorable null

The least favorable null configurations (LFC) identify the “worst case scenarios” that lead to the maximum type 1 error rate over the full parameter space. Note that the test statistics for evaluating H0, T, is stochastically bounded by both and . Thus is maximum when equality holds, i.e. when T attains one of the bounds. For evaluating the single hypothesis H0, the above situation arises when the combined mean μ is equal to one of the monotherapies and infinitely larger than the other monotherapy. Hence, we conclude that the LFC for H0 occurs under such situations. This LFC can be best formulated as where and . Here a ≫ b indicates that a is infinitely larger than b. Following from here, the LFC for H0 in equation (3) occurs when the mean response μ under each combination (i, j) is equal to one of the monotherapies and infinitely larger than the other. This can be written as where and τ ∈ {A, B} means that, τ is a map from K to {A, B}, i.e. We will call τ a configuration of A's and B's. Now, some τ ∈ {A, B} are infeasible in the sense that The set of all infeasible τ is given by where and τ is a map from to . As one of the reviewers have pointed out, it might be interesting to explore the size of the infeasible LFC as compared to the set of feasible ones. We have written a R code to visualize the cardinality of for different choices of dose levels of Drug A (r) and dose levels of Drug B (s). It is added to the supplementary material. Table 1 shows the cardinality of for the different choices of dose levels of Drug A and Drug B. It is evident from the table that the rate of increase of infeasible LFC is very high as the number of dose levels increases. With the help of an illustrative example, we have shown later the set of infeasible LFC (2 out of 16) in a 3 × 3 drug combination trial, i.e. where r = 2 and s = 2. Furthermore, since we are interested in drug combination studies for Phase II clinical trials, the dose levels of the two drugs are unlikely to go beyond 4 or 5.
Table 1.

Cardinality of infeasible LFC and the cardinality of all the LFC (in bracket) for different choices of dose levels of Drug A and Drug B, r and s, respectively.

s
r1234
10 (2)0 (4)0 (8)0 (16)
20 (4)2 (16)18 (64)110 (256)
30 (8)18 (64)282 (512)3030 (4096)
40 (16)110 (256)3030 (4096)58634 (65536)
Cardinality of infeasible LFC and the cardinality of all the LFC (in bracket) for different choices of dose levels of Drug A and Drug B, r and s, respectively. Consider the critical value C for the above approach such that it satisfies We reject the global null in equation (3) when the observed T > C. Furthermore, all the component test statistics T are tested against the critical value C and decisions are taken following a single-step testing procedure.[21] The adjusted p-value for each component hypothesis H0 for the LFC approach can be obtained by computing the following probability; , where t is the observed value of the test statistics T in equation (4). It is interesting to note in Table 1 that for a 5 × 5 combination trial, where r = 4 and s = 4 approximately 90% of the observed LFC are infeasible. One needs to compute the type 1 error only under 10% LFC and type 1 error computation is less time consuming under such a scenario as compared to computing the type 1 error under all possible LFC. The “Max” test together with the earlier mentioned subset pivotality property[22] ensures that the FWER is controlled in the strong sense in the above testing approach. The following example shows an illustration of how one can obtain the maximum type 1 error using the LFC approach with four active drug combinations.

3.1.2 Example

We consider a 3 × 3 drug combination study, i.e. a study with two drugs and two active doses per drug. Then , where is the test statistics for comparing the ij drug combination with its monotherapies. The formal set of LFC is given in Table 2. In Table 2, LF1 denotes the set of dose combinations where the dose response means are equal to their first monotherapies and infinitely larger than their second monotherapies. LF7 and LF10 are marked grey because they are infeasible. The reason why LF7 is infeasible is that,
Table 2.

All least favorable null configurations for a 3 × 3 drug combination trial with four dose combinations.

Least Favorable Null(1.1)(1.2)(2.1)(2.2)
LF 1 LFC11A LFC12A LFC21A LFC22A
LF 2 LFC11A LFC12A LFC21A LFC22B
LF 3 LFC11A LFC12A LFC21B LFC22A
LF 4 LFC11A LFC12A LFC21B LFC22B
LF 5 LFC11A LFC12B LFC21A LFC22B
LF 6 LFC11A LFC12B LFC21A LFC22B
LF 7 LFC11A LFC12B LFC21B LFC22A
LF 8 LFC11A LFC12B LFC21B LFC22B
LF 9 LFC11B LFC12A LFC21A LFC22A
LF 10 LFC11B LFC12A LFC21A LFC22B
LF 11 LFC11B LFC12A LFC21B LFC22A
LF 12 LFC11B LFC12A LFC21B LFC22B
LF 13 LFC11B LFC12B LFC21A LFC22A
LF 14 LFC11B LFC12B LFC21A LFC22B
LF 15 LFC11B LFC12B LFC21B LFC22A
LF 16 LFC11B LFC12B LFC21B LFC22B
All least favorable null configurations for a 3 × 3 drug combination trial with four dose combinations. , which gives a contradiction. Similarly LF10 leads to another contradiction and thus it is infeasible as well. The maximum type 1 error can be computed as where each probability within the max is attained under each LFC in Table 2. Obviously the FWER is bounded by the maximum over all the LF in Table 2, but note that we have omitted the infeasible LF7 and LF10. We will now show that the FWER under all mean constellations are controlled by the feasible LFC (LFC obtained after omiting LF7 and LF10). Clearly, the FWER under any mean constellation increases when replacing inequalities μ ≤ μ and μ ≤ μ0 by equalities. The resulting FWER is then dominated by the LF in Table 2 with similar equalities, i.e. the FWER under any mean constellation with μ ≤ μ is bounded by the corresponding FWER under the LFC where μ = μ (LFC where τ = A), and the FWER under any mean constellation with μ ≤ μ0 is bounded by the corresponding FWER under the LFC where μ = μ0 (LFC where τ = B). Assume now that μ corresponds to the set similar to LF7 where the dose combination means are equal to one monotherapy and finitely larger than the other monotherapy. We call this case Since we are omitting LF7 which is a bound for the FWER under the mean constellation F, we explore other bounds for this within the FWER under the feasible LF's. Note that this particular configuration, F, can hold true only when all the inequalities are substituted by equality, i.e. μ = μ = μ0 for all (i, j). Under this configuration, the FWER is bounded by the FWER of any feasible LF because the probability of rejection goes up when one of the means is really high. Hence controlling the maximum type 1 error within the significance level α ensures that the FWER is controlled in the strong sense. Bootstrap test: Under this approach, we aim to approximate the true null distribution of the test statistics using bootstrap methods. An illustration of the null parameter space is given in Figure 1. Here we estimate the parameters either under the null boundary constraint or under the space of all null configurations including the null interior constraint shown in Figure 1. With the bootstrap method, we generate samples under model (1) using estimates of the parameter μ and σ[2]. These bootstrap samples are then utilized to obtain a sample of bootstrap test statistics with an empirical distribution, that is ideally a good approximation of the unknown underlying true null distribution of the test statistics T in equation (4). The parametric bootstrap approach can be outlined in more details by the following steps. Note that 2a and 2b below show the two options of projecting on the null space corresponding to the two different constraints illustrated in Figure 1.
Figure 1.

The null parameter space for any dose combination.

For a given multiple dose combination study, compute the test statistics where and are defined earlier in section 3. For the given dataset estimate the mean under the constraints, i.e. a) such that, b) such that, The standard deviation is estimated as , where is the unrestricted maximum likelihood estimate (mle) of μ. (a) Simulate 5000 normal distributed random variables with mean estimate and (b) Simulate 5000 normal distributed random variables with mean estimates from the earlier step. The standard deviation estimates remain the same in both the cases. For each simulated data compute the test statistics T in equation (4). Find out the proportion of times the test statistics from the simulated data is greater than the observed test statistics both in case (a) and (b). This gives our p-value for the above bootstrap test under option (a) and (b), respectively. The null parameter space for any dose combination. Note that instead of calculating a p-value we could equivalently calculate the (1 – α) quantile of the bootstrap distribution for T and use this as the critical value for T. We will refer to this bootstrap critical value in the next section. We have used the unrestricted mle to estimate in the above bootstrap approach instead of the restricted estimate in order to make sure that the bootstrapped test statistics satisfy the subset pivotality criteria mentioned earlier. This ensures in the above “Max” test, that the FWER is strongly controlled under the bootstrap approach as well.[20] By this, we can not only claim that the drug-combination is beneficial overall but also provide the set of dose combinations showing beneficial effect in our study. We have observed in simulation studies that the option 2a provides better type 1 error control than option 2b. This is not surprising because the null hypotheses on the boundary are less favorable than those in the interior. In the numerical example and the simulation studies below, we present only the bootstrap approach under the option 2a.

3.2 Relationship between the different approaches

With the bootstrap approach, the critical value depends on the data via the constraint parameter estimates. Hence, the critical value is a random variable in the bootstrap approach whereas it is a fixed number for the LFC and Hung's approach.[7] In this section, we investigate different approaches by comparing their critical values under a particular dose combination setup. We illustrate the distribution of critical values for the different methods in Figure 2. For simplicity we simulate data from a balanced factorial design with r = 1, s = 1, and n = 25 for all (i, j) combinations. Further we fix the first effect size at 0, i.e. set δ1 = μ11 – max(μ10, μ01) = 0 and the second effect size (δ2 = μ12 – max(μ10, μ02)) begins with 0 and increases along the X-axis. For the above setup, we plot the critical value of the test statistic (along the Y-axis) for the LFC approach, Hung's[7] approach, Bootstrap approach and the oracle critical value, i.e. the critical value that can be derived if we know the unknown true null distribution of the test statistics. Since there is only one combination involved, computing the distribution of the test statistic under the global null H0 in equation (3) is not complicated. We observe that the critical value under the LFC is indeed a limiting case of the oracle critical values.
Figure 2.

Distribution of critical value for the different methods under the set up: Drug A and Drug B both have one active dose group with sample size per dose group=25. The first effect size (δ1) is 0 and the second effect size (δ2) varies. Note that is addressed as in the article and the plots are shown in terms of the non-centrality parameter of test statistics , i.e. , which is the second effect size scaled by the harmonic sum of sample size of the dose combination and the second monotherapy. (a) Boxplot distribution at δ211 = 0.707; (b) boxplot distribution at δ211 = 1.591; (c) boxplot distribution at δ211 = 3.359; (d) boxplot distribution at δ211 = 5.127.

Distribution of critical value for the different methods under the set up: Drug A and Drug B both have one active dose group with sample size per dose group=25. The first effect size (δ1) is 0 and the second effect size (δ2) varies. Note that is addressed as in the article and the plots are shown in terms of the non-centrality parameter of test statistics , i.e. , which is the second effect size scaled by the harmonic sum of sample size of the dose combination and the second monotherapy. (a) Boxplot distribution at δ211 = 0.707; (b) boxplot distribution at δ211 = 1.591; (c) boxplot distribution at δ211 = 3.359; (d) boxplot distribution at δ211 = 5.127. Our LFC approach is similar to the one in Hung's[7] approach with the exception that it does not rely on an asymptotic normal approximation. The LFC approach identifies the same configurations as the extreme parameter configurations of the parameter δ (where ) introduced by Hung.[7] Under these extreme configurations using multiple testing theory, we show that the LFC leads to a multivariate t distribution for the test statistics T in equation (4) under the null H0 in equation (3). It is likely that the asymptotic approximation was introduced by Hung[7] because numerical techniques for efficient evaluation of probabilities over rectangular region[23] were not available then. Nevertheless, our illustration for a 2 × 2 factorial design in Figure 2 shows that the difference in the critical value between the two methods is marginal, with the LFC being more conservative than Hung's approach. We have conducted 1000 simulations at some particular fixed values of δ1 and δ2. δ1 is fixed at 0 and 4 different values of δ2 are selected and the boxplots distribution of the critical values for the bootstrap method under the different values of δ2 is shown in Figure 2. As one can see from the plot, the critical value under the bootstrap method is approximately centered around the oracle critical value at each pre-selected δ2 (shown by the dotted green line). However, there might be cases where the critical value under the bootstrap method is smaller or larger compared to the oracle critical value. Asymptotically the type error should be controlled strongly with the subset pivotality criteria[20] under the bootstrap approach but the critical value is sometimes underestimated as shown in Figure 2(c), thereafter explaining the inflation in type 1 errors observed in our simulation scenarios shown in Table 15.
Table 15.

Empirical type 1 error rate for the different methods: Scenario 2.

Sample SizeBonfHungBootLFC
100.0380.0500.0580.038
250.0490.0530.0610.049
500.0470.0490.0500.047
750.0430.0450.0470.044
1000.0480.0500.0490.048
A similar illustration of the above scenario with δ1≠0 is given in Figure 3 of Appendix 1. In summary, it is observed that the critical value under the bootstrap method is always centered around the least favorable null critical value and it is with a high probability below the least favorable critical value for smaller δ2. This indicates that the bootstrap method will give more power than the LFC approach. From the above illustration, it is expected that the LFC approach gives more conservative results compared to the bootstrap approach and it also shows that Hung's[7] approach behaves similar to the LFC approach.

4 Numerical example

We consider here an example from Hung[7] to illustrate the methods discussed in the previous sections. A combination of a diuretic (drug B) and an ACE inhibitor (drug A) is tested for the efficacy in reduction of sitting diastolic pressure (SiDBP) with a pooled standard deviation of σ = 7.07. The response means and sample sizes are summarized in Table 3.
Table 3.

Mean responses and the sample sizes (in bracket) of the drug combination study.

Drug B
Drug A012
00 (75)1.8 (74)2.8 (48)
11.4 (75)2.8 (75)4.5 (50)
22.7 (74)5.7 (74)7.2 (48)
34.6 (48)8.2 (49)10.9 (48))
Mean responses and the sample sizes (in bracket) of the drug combination study. In total 750 patients are randomized to receive one of the 12 dose combinations. The primary objective of the study is to test whether there exist at least a combination (i, j) which is superior to the component drugs. Table 4 shows the unadjusted p-values as well as the adjusted p-values from the different methods for each dose combination.
Table 4.

Unadjusted and adjusted p-values for each drug combination, when different methods are applied to the data example in Table 3.

Dose CombTStatUnadjPBonfPBootPLFCP
(1, 1)0.8630.1941.0000.6500.709
(1, 2)1.1900.1170.7030.4520.512
(3, 1)2.5070.0060.0370.0240.036
(2, 1)2.5810.0050.0300.0200.029
(2, 2)3.0490.0010.0070.0040.007
(3, 2)4.3650.0000.0000.0000.000

Dose Comb: Different dose combinations TStat : Test statistics Tij testing for H0ij. UnadjP is the one-sided raw p-value testing H0ij against H1ij. BonfP, BootP and LFCP are the one sided Bonferroni adjusted, Bootstrap adjusted and LFC adjusted p-values respectively.

Unadjusted and adjusted p-values for each drug combination, when different methods are applied to the data example in Table 3. Dose Comb: Different dose combinations TStat : Test statistics Tij testing for H0ij. UnadjP is the one-sided raw p-value testing H0ij against H1ij. BonfP, BootP and LFCP are the one sided Bonferroni adjusted, Bootstrap adjusted and LFC adjusted p-values respectively. We see from Table 4 that all the approaches proposed by us suggest that at 5% significance level the combinations (2, 1), (2, 2), (3, 1), and (3, 2) are superior to the monotherapies. For the Bonferroni adjusted method, the p-values are compared to the local significance level 0.8% to control the overall type 1 error at 5%. This numerical example was also investigated by other authors. Based on a global test, Hung[7] concluded that there exists at least one combination that is superior to the monotherapies. He also applied an approximation of James[24] to adjust the p-values and thereby identified the dose combination (2, 1), (2, 2) and (3, 2) as superior. But now that there is no need to approximate the distribution of the test statistics because Genz and Bretz[23] proposed numerical techniques for an efficient evaluation of multivariate t-distribution probabilities over rectangular regions. Hellmich and Lehmacher[8] implemented the AVE and MAX test proposed by Hung[7] using the above proposed adjustment on the same dataset and concluded that the combination (2, 2) and (3, 2) are superior at 2.5% significance level (one-sided) with strong FWER control. Furthermore, using Holm's method, they showed that dose combinations (2, 1) and (3, 1) are also effective at significance level 2.5% (one-sided) to control the FWER at 5%. Soulakova[12] cited the same example and concluded using Holm's approach that all the four combinations: (2, 1), (2, 2), (3, 1), and (3, 2) are superior and effective at significance level 5%. However, unlike Hellmich and Lehmacher,[8] they conducted a test for both effectiveness and superiority and did not assume a priori that the dose combinations means are always greater than or equal to the monotherapy means.

5 Simulation studies

In this section, a simulation study is presented. The main objective of this simulation study is to compare the performance of the following approaches: (1) Hung's method, (2) parametric bootstrap method, (3) Bonferroni correction and (5) LFC approach with regards to their (a) ability to control the overall type 1 error at 5% significance level and (b) power to detect the superior dose combinations. We consider 11 scenarios overall, amongst which Scenario 1 and Scenario 2 are designed to investigate the strong control of type 1 error rate and Scenario 3 to Scenario 11 are designed to compare power performance of the different testing strategies. In Scenario 1 to Scenario 3 we are considering a balanced factorial design with r = 2, s = 1, and n = n for all (i, j) combinations. In Scenario 4 to Scenario 11 we are considering an unbalanced factorial design with r = 3, s = 2, and differing n for the (i, j) combinations.

5.1 Simulation scenarios

The simulation scenarios are divided into two parts. Section 5.1.1 refers to some new scenarios which are introduced in this article and Section 5.1.2 refers to the scenarios taken from Hung (2000).[7]

5.1.1 New scenarios

We are considering three scenarios. Scenario 1 refers to an extreme situation where the parameters are drawn from the restricted parameter space LFCfeasible. Note that in Table 5 where we are presenting Scenario 1, we assign δ = 2 and a = 9999, where 9999 represents an essentially large number close to the LFC where some dose combination means are infinitely larger than those of the monotherapies. Scenario 1 is to evaluate the ability of the different methods in controlling the type 1 error rate under extreme situations. For this we simulate the data randomly from one of the four cases shown in Table 5, where each LF represents a configuration that can occur in a 3 × 2 factorial design:
Table 5.

Four least favorable null configurations in a 3 × 2 factorial design.

Least Favorable ConfigDrug ADrug B
01
LF 4 0 δ δ + a
1 δ δ + a
2 δ δ + a
LF 3 0 δ δ + a
1 δ δ + a
2δ + 2aδ + 2a
LF 2 0 δ δ + a
1δ + 2aδ + 2a
2 δ δ + a
LF 1 0 δ δ
1δ + aδ + a
2δ + aδ + a
Four least favorable null configurations in a 3 × 2 factorial design. However, as Scenario 1 is very extreme, Scenario 2 is considered to evaluate the performance of the different methods under a more realistic set up, where the value of a in Scenario 1 is replaced by 0.7. To evaluate the power performance across different sample sizes in a balanced design, data are simulated under Scenario 3 shown in Table 6. We consider five different sample sizes for the above described scenarios: 10, 25, 50, 75, 100. We evaluate the empirical type 1 error and power based on 5000 simulation runs assuming normally distributed errors with standard deviation 1. For the parametric bootstrap method, 5000 bootstrap samples are used.

5.1.2 Scenarios from Hung (2000)

Scenarios 4–11 are reproduced from Hung (2000).[7] They are introduced to assess the power performances of the different methods under different effect sizes. Effect sizes here indicate the value of the contrast comparing the dose combination with the best monotherapy (θ = μ – max(μ, μ0)). We want to further investigate the power performance of the different methods under a balanced and unbalanced design. Two possible effect sizes (E1 and E2) are shown in Tables 7 and 8. The average effect size (average of θ) is 0.3 for both designs but in E1 all effect sizes is equal and in E2 the combination of the lower dose A1 has a smaller effect compared to E1. For the above two designs, we consider four possible sample size allocations (S1, S2, S3, S4) given in Tables 9 to 12. S1 is a balanced design. S2 is introduced to increase the power of dose combination (A2, B1). Here (A2, B1) denotes the dose combination with dose level 2 of Drug A and dose level 1 of Drug B. S3 is designed such that more sample size is allocated to monotherapies corresponding to Drug A. This is introduced to ensure sufficient power of the drug combinations when compared with the first monotherapies. S4 is introduced to ensure more power for combinations with higher doses, particularly when the data is simulated from E2.
Table 7.

Dose-response means for the factorial design E1.

Drug B
Drug AB0B1B2
A000.20.5
A10.10.50.8
A20.30.60.8
A30.60.90.9
Table 8.

Dose-response means for the factorial design E2.

Drug B
Drug AB0B1B2
A000.20.5
A10.10.250.65
A20.30.700.90
A30.61.01.0
Table 9.

Sample size scenario (S1) for the drug combination designs (E1 and E2).

Drug B
Drug AB0B1B2
A0505050
A1505050
A2505050
A3505050
Table 12.

Sample size scenario (S4) for the drug combination designs (E1 and E2).

Drug B
Drug AB0B1B2
A0505656
A1303030
A2585858
A3585858
Scenario 3: A balanced design scenario devised to evaluate the power of the different methods across different sample sizes. Dose-response means for the factorial design E1. Dose-response means for the factorial design E2. Sample size scenario (S1) for the drug combination designs (E1 and E2).

5.2 Simulation results

Tables 14 and 15 shows how the type 1 error rate is controlled under the different methods for Scenario 1 and Scenario 2, respectively. Table 16 presents the empirical power of the different approaches under Scenario 3. Table 17 presents the empirical power of the different approaches under each effect size pattern and each sample size allocation for scenarios in Table 13.
Table 14.

Empirical type 1 error rate for the different methods: Scenario 1.

Sample SizeBonfHungBootLFC
100.0450.0570.0470.045
250.0500.0550.0530.051
500.0470.0490.0480.047
750.0440.0450.0460.044
1000.0480.0500.0490.049
Table 16.

Empirical power of the 5% level max test for the different methods under Scenario 3.

Sample SizeBonfHungBootLFC
100.14880.17180.21920.1530
250.43300.44320.51040.4380
500.78860.79360.83580.7904
750.92880.93080.94500.9304
1000.98000.97900.98400.9806
Table 17.

Empirical power of the 5% level max test for the different methods under the data scenarios in Table 13.

ScenarioDose Response DesignSample SizeBonfHungBootLFC
Scenario 4S10.56220.58180.65960.5690
Scenario 5E1S20.56260.57500.64660.5670
Scenario 6S30.41600.45100.54780.4222
Scenario 7

S4
0.5794
0.5974
0.6702
0.5846
Scenario 8S10.72140.73500.79680.7286
Scenario 9E2S20.75380.77500.82540.7570
Scenario 10S30.61020.63620.73220.6154
Scenario 11S40.79300.80300.85720.7966
Table 13.

Scenario 4–Scenario 11.

ScenarioDose Response DesignSample Size
Scenario 4E1S1
Scenario 5E1S2
Scenario 6E1S3
Scenario 7E1S4
Scenario 8E2S1
Scenario 9E2S2
Scenario 10E2S3
Scenario 11E2S4
Sample size scenario (S2) for the drug combination designs (E1 and E2). Sample size scenario (S3) for the drug combination designs (E1 and E2). Sample size scenario (S4) for the drug combination designs (E1 and E2). Scenario 4–Scenario 11. Empirical type 1 error rate for the different methods: Scenario 1. Empirical type 1 error rate for the different methods: Scenario 2. Empirical power of the 5% level max test for the different methods under Scenario 3. Empirical power of the 5% level max test for the different methods under the data scenarios in Table 13. From Tables 14 and 15 we observe that the Bonferroni method and the LFC approach show a more conservative behavior compared to Hung's and the bootstrap approach. Note that though the Bonferroni method is criticized quite often in the literature, it performs almost as good as the LFC approach in our simulations. The type 1 error is somewhat inflated for small sample sizes (e.g. n = 10, 25 ∀i, j) with the bootstrap method. Tables 16 and 17 indicate that the parametric bootstrap approach shows uniformly better power performance than the other methods across all the sample sizes. This is because, under the alternative scenarios, the critical value of the bootstrap method is mostly below the critical value of the LFC approach. This has been elaborated in Section 4. The power performance of Hung's[7] approach is similar to the power performance of the LFC approach. This is because they are essentially the same, one using an approximate and the other the exact distribution of the same test statistics. Note that we have used only 5000 iterations for the bootstrap approach in our simulation studies but as the number of iteration increases the inflation in type error is expected to reduce. Nevertheless, there will still be some inflation regardless of the number of bootstrap samplings and as the number of bootstrap iteration increases the method becomes more time consuming and the improvement is not significant, so we adhered to 5000 iterations. Summing up, the bootstrap approach is giving power improvement by 6–10% over the other approaches across all the scenarios. We further observe from Table 17 that we can gain in power by choosing an unbalanced design with more sample size allocation in the combination doses compared to the component doses. In S3 the power performance dropped as more sample size is allocated to the monotherapies of Drug A compared to the dose combinations, whereas in S4, the power across all the methods are becoming better when more sample size is allocated to higher dose combinations compared to the monotherapies. The marginal improvement of Hung's method over the LFC approach is due to the approximate nature of the first.

6 Discussion

We observe from our simulation experiments that both the bootstrap method and the LFC approach, proposed in this article, meet the nominal level for attaining the global null hypothesis if per group sample sizes are not too small (e.g. 50 or more). They also strongly control the FWER at the desired significance level. Since the LFC approach is conservative across all sample sizes, the bootstrap approach is providing more power compared to the LFC. The LFC and Bonferroni approach performs surprisingly similar to each other. This is likely due to the fact that the least favorable null configuration will be the one where the test statistics are either independent or have low correlation. This follows from Slepian's inequality[25] which says that the type 1 error from a “max” test, like the ones discussed in this article, increases with the decrease in the correlation between the component test statistics that combine to generate the max test statistic. The ideal maxima will occur when all the test statistics are uncorrelated, which will lead to the similar type 1 error as the type 1 error under the Bonferroni adjustment. However, in reality when there are multiple dose combinations involved, it is improbable to achieve this ideal maxima and hence the LFC type 1 error will always be greater than or equal to the Bonferroni type 1 error. This is also evident in our simulations. Note that we have only tested for superiority in our null hypothesis in equation (3). In order to test for effectiveness, we only need to add one additional test per dose combination, namely the test against placebo, such that we have overall the union of an intersection of three (instead of two) tests. Moreover, it is unlikely that we have non-efficacy but superiority with regard to both monotherapies. Hence, we have only considered superiority in our testing problem. Resampling-based bootstrap approaches have already been suggested by other authors.[15,26] Soulakova[15] proposed a resampling based framework for a multiple dose factorial design where she identified all the effective and superior combinations without any considerations of the role of the nuisance parameters (difference in monotherapies) in the resampling distribution. Accordingly, she observed family-wise error rate inflations in multiple situations. Frommolt and Hellmich[26] addressed this issue by a resampling-based bootstrap approach, where the nuisance parameters are estimated and accounted for in the resampling approximation of the test statistic's null distribution. But this bootstrap-based testing performed similar to the Hung's approach. The virtue of the parametric bootstrap approach suggested in this article is that, unlike the earlier approaches, it is performing better than the Hung's[7] approach under the alternative hypothesis. As one of the reviewers pointed out, a potential downside of the parametric bootstrap approach is that it relies on the normal distribution assumption, which may not be always true. However, it is only a concern for small sample sizes; for large sample sizes, one can easily show using the central limit theorem that the test statistics is approximately t distributed regardless of the underlying data distribution. The methods suggested here only provide a set of superior dose combinations but do not propose an optimal dose for future use in the drug developments process. We also cannot infer anything beyond the observed doses if the nature of dose response relationship is not known a priori. Hence, there is a strong interest in estimating the dose–response relationship for the drug combination. Hung[5,27] suggested a response surface methodology approach, where after attaining global superiority one can utilize the biological information on the drug combination study and apply a statistical model to estimate the relationship between the drug dosages and mean response. This approach helps us to obtain an optimal dose and make inference around this optimal dose in a multiple dose drug combination trial. But often it happens that the true dose–response pattern is not known and then choosing an appropriate dose response model becomes difficult. Hellmich and Lehmacher[8] proposed a closed testing procedure for estimating the minimum effective dose and highest effective dose levels in a dose–response bi-factorial design but they had to assume monotonicity properties to obtain the likelihood ratio tests and multiple contrast tests for their proposed hypotheses. To the best of our knowledge, there exists no approach where one can simultaneously control the FWER and infer on the dose–response relationship in a multiple dose drug combination study without the monotonicity assumption. Hence, it might be interesting to extend our bootstrap-based multiple testing approach to a modelling framework. Distribution of critical value for the different methods under the set-up: Drug A and Drug B both have one active dose group with sample size per dose group = 25. The first effect size (δ1) is 0.5 and the second effect size (δ2) varies along X-axis. Note that is addressed as in the article and the plots are shown in terms of the non-centrality parameter of test statistics , i.e. , which is the second effect size scaled by the harmonic sum of sample size of the dose combination and the second monotherapy. This plot is provided to analyze the empirical power performance of the different methods. (a) Boxplot distribution at δ211 = 0.707; (b) boxplot distribution at δ211 = 1.591; (c) boxplot distribution at δ211 = 3.359; (d) boxplot distribution at δ211 = 5.127. A multiple dose drug combination factorial design. : The mean constellation matrix following from . Click here for additional data file. Supplemental Material for Testing multiple dose combinations in clinical trials by Saswati Saha, Werner Brannath and Björn Bornkamp in Statistical Methods in Medical Research
Table 6.

Scenario 3: A balanced design scenario devised to evaluate the power of the different methods across different sample sizes.

Drug B
Drug A01
022
122.5
222.5
Table 10.

Sample size scenario (S2) for the drug combination designs (E1 and E2).

Drug B
Drug AB0B1B2
A0509035
A1353535
A2909035
A3353535
Table 11.

Sample size scenario (S3) for the drug combination designs (E1 and E2).

Drug B
Drug AB0B1B2
A0502020
A1705050
A2705050
A3705050
Table 18.

A multiple dose drug combination factorial design.

Drug B
Drug A012..s
0 μ 00 μ 01 μ 02 .. μ 0s
1 μ 10 μ 11 μ 12 .. μ 1s
.......
r μ r0 μ r1 μ r2 .. μ rs
Table 19.

: The mean constellation matrix following from .

Drug B
Drug A12. s 2 . s 1 . s
1 μ 10 μ 10 ... μ 10 μ0s1+1 μ 0s
2 μ 20 μ 20 . μ 20 μ0s2+1 .. μ 0s
........
r........
12.. s 1 .. s
1AA..ABBB
.AA.BBBBB
.........
.........
r........
  14 in total

1.  Resampling-based and other multiple testing strategies with application to combination drug trials with factorial designs.

Authors:  Julia N Soulakova
Journal:  Stat Methods Med Res       Date:  2010-07-28       Impact factor: 3.021

2.  Closure procedures for monotone bi-factorial dose-response designs.

Authors:  M Hellmich; W Lehmacher
Journal:  Biometrics       Date:  2005-03       Impact factor: 2.571

3.  Regulatory challenges for new formulations of controlled substances in today's environment.

Authors:  Cynthia G McCormick
Journal:  Drug Alcohol Depend       Date:  2006-03-29       Impact factor: 4.492

4.  Approximate multinormal probabilities applied to correlated multiple endpoints in clinical trials.

Authors:  S James
Journal:  Stat Med       Date:  1991-07       Impact factor: 2.373

5.  Comparison of several testing strategies for combination drug efficacy trials based on the closure principle.

Authors:  Julia N Soulakova
Journal:  Stat Med       Date:  2009-01-30       Impact factor: 2.373

6.  Multiple test procedures for dose finding.

Authors:  A C Tamhane; Y Hochberg; C W Dunnett
Journal:  Biometrics       Date:  1996-03       Impact factor: 2.571

7.  Testing whether an identified treatment is best.

Authors:  E M Laska; M J Meisner
Journal:  Biometrics       Date:  1989-12       Impact factor: 2.571

8.  Evaluating the efficacy of a combination therapy.

Authors:  S M Snapinn
Journal:  Stat Med       Date:  1987-09       Impact factor: 2.373

9.  Testing for the existence of a desirable dose combination.

Authors:  H M Hung; G Y Chi; R J Lipicky
Journal:  Biometrics       Date:  1993-03       Impact factor: 2.571

10.  On Identifying Minimum Efficacious Doses in Combination Drug Trials.

Authors:  Julia N Soulakova; Allan R Sampson
Journal:  Stat Biopharm Res       Date:  2009       Impact factor: 1.452

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.