| Literature DB >> 35102685 |
Elias Laurin Meyer1, Peter Mesenbrink2, Cornelia Dunger-Baldauf3, Ekkehard Glimm3,4, Yuhan Li2, Franz König1.
Abstract
Platform trials have become increasingly popular for drug development programs, attracting interest from statisticians, clinicians and regulatory agencies. Many statistical questions related to designing platform trials-such as the impact of decision rules, sharing of information across cohorts, and allocation ratios on operating characteristics and error rates-remain unanswered. In many platform trials, the definition of error rates is not straightforward as classical error rate concepts are not applicable. For an open-entry, exploratory platform trial design comparing combination therapies to the respective monotherapies and standard-of-care, we define a set of error rates and operating characteristics and then use these to compare a set of design parameters under a range of simulation assumptions. When setting up the simulations, we aimed for realistic trial trajectories, such that for example, a priori we do not know the exact number of treatments that will be included over time in a specific simulation run as this follows a stochastic mechanism. Our results indicate that the method of data sharing, exact specification of decision rules and a priori assumptions regarding the treatment efficacy all strongly contribute to the operating characteristics of the platform trial. Furthermore, different operating characteristics might be of importance to different stakeholders. Together with the potential flexibility and complexity of a platform trial, which also impact the achieved operating characteristics via, for example, the degree of efficiency of data sharing this implies that utmost care needs to be given to evaluation of different assumptions and design parameters at the design stage.Entities:
Keywords: adaptive trials; clinical trial simulation; combination therapy; platform trials
Mesh:
Year: 2022 PMID: 35102685 PMCID: PMC9304586 DOI: 10.1002/pst.2194
Source DB: PubMed Journal: Pharm Stat ISSN: 1539-1604 Impact factor: 1.234
FIGURE 1Schematic overview of the proposed platform trial design. New cohorts consisting of a combination therapy arm, a monotherapy arm using the same compound in every cohort (referred to as “backbone monotherapy”), an add‐on monotherapy arm which is different in every cohort and a SoC arm are entering the platform over time. While the add‐on monotherapy and therefore the combination therapy is different in every cohort (as indicated by the differently shaded colors), the backbone monotherapy and SoC are the same in every cohort (as indicated by the same colors). Each cohort has an interim analysis after about half of the initially planned sample size, after which the cohort can be stopped for early efficacy or futility. The red brackets indicate the testing strategy within each cohort, that is, comparison of combination therapy against both monotherapies and both monotherapies against SoC. We differentiate between per‐cohort and per‐platform operating characteristics (OCs)
FIGURE 2Schematic overview of the different levels of sharing. No sharing happens if only “cohort” data are used. If sharing “all” data, whenever in any cohort an interim or final analysis is performed, all SoC and backbone monotherapy data available from all cohorts are used. If sharing only “concurrent” data, whenever in any cohort an interim or final analysis is performed, all SoC and backbone monotherapy data that was collected during the active enrollment time of the cohort under investigation are used. If sharing “dynamically,” whenever in any cohort an interim or final analysis is performed, the degree of data sharing of SoC and backbone monotherapy data from other cohorts increases with the homogeneity of the observed response rate of the respective arms. A solid fill represents using data 1‐to‐1, while a dashed fill represents using discounted data (for more information see Appendix A.1). If at any given time there are k active cohorts, the allocation ratio is 1:1:1:1 in case of no data sharing and k:k: 1:1 otherwise (combination: add‐on monotherapy: backbone monotherapy: SoC). This allocation ratio is updated for all active cohorts every time the number of active cohorts k changes due to dropping or adding a new cohort
Operating characteristics used in this paper and their definitions
| Name | Definition |
|---|---|
| PCP | “Per‐Cohort‐Power,” the ratio of the sum of true positives among the sum of truly efficacious cohorts (i.e., the sum of true positives and false negatives) across all platform trial simulations, that is, the probability of a true positive decision for any new cohort entering the trial. This is a measure of how wasteful the trial is with superior therapies. |
| PCT1ER | “Per‐Cohort‐Type‐1‐Error,” the ratio of the sum of false positives among the sum of all truly not efficacious cohorts across all platform trial simulations, that is, the probability of a false positive decision for any new cohort entering the trial. This is a measure of how sensitive the trial is in detecting futile therapies. |
| FWER | “Family‐wise type 1 Error Rate”, the proportion of platform trials, in which at least one false positive decision has been made (i.e., probability of at least one false positive decision across all cohorts), where only such trials are considered, which contain at least one cohort that is in truth futile. Formal definition: |
| FWER BA | “Family‐wise type 1 Error Rate Bayesian Average”, the proportion of platform trials, in which at least one false positive decision has been made (i.e., probability of at least one false positive decision across all cohorts), regardless of whether or not any cohorts which are in truth futile exist in these trials. Formal definition: |
| Disj Power | “Disjunctive Power”, the proportion of platform trials, in which at least one correct positive decision has been made (i.e., probability of at least one true positive decision across all cohorts), where only such trials are considered, which contain at least one cohort that is in truth superior. Formal definition: |
| Disj Power BA | “Disjunctive Power Bayesian Average”, the proportion of platform trials, in which at least one correct positive decision has been made (i.e., probability of at least one true positive decision across all cohorts), regardless of whether or not any cohorts which are in truth superior exist in these trials. Formal definition: |
Simulation Setup Overview. For different simulation parameters, we differentiate between parameters that are considered a design choice and parameters that are considered an assumption regarding the future course of the platform or treatment effects. For the investigated values, we state in bold which value was considered the default value, that is, unless stated otherwise in a particular figure, the parameters was set to this value. For some parameters there is no default (e.g., when shown as a simulation dimension in every figure or if chosen as a fixed design parameter in section 2)
| Name | Type | Investigated values | Description |
|---|---|---|---|
| Maximum number of cohorts | Assumption | 3, 5, | Assumed maximum number of cohorts per platform (can be less in individual simulations) |
| Cohort inclusion rate | Assumption | 0.01, | Probability to include a new cohort in the ongoing platform trial after every simulated patient (unless maximum number of cohorts is reached). The default value leads to reaching the assumed maximum number of cohorts in nearly every simulation for nearly every sample size. |
| Treatment efficacy setting | Assumption |
| Assumed treatment effects of the different study arms. For more information, see Table |
| Final cohort sample size | Design choice | 100, 200, 300, 400, | Number of patients after which final analysis in a cohort is conducted (interim analysis always after half the final sample size) |
| Data sharing | Design choice | All, concurrent, dynamic, cohort | Different methods of data sharing used at analyses, ranging from full pooling to not sharing at all. For more information, see section |
| Allocation ratios | Design choice | Balanced or unbalanced (depends on data sharing) | Allocation ratio to treatment arms within a cohort. Depending on the method of data sharing, the allocation ratio is set to either balanced or unbalanced (i.e., randomizing more patients to combination and add‐on monotherapy). For more information, see section |
| Bayesian decision rule | Design choice |
| Thresholds used in the Bayesian decision making at interim and final. Values other than the default values are used only in Figure |
| Interim analysis | Design choice |
| Binding rules on whether or not a cohort can be stopped at interim for early futility (note: cohorts can always stop for early efficacy). For more information, see section |
Overview of different treatment effect assumptions. The priors T , T and T for γ , γ and γ as described in section A.3 are all pointwise with a support of 1,2 or 3 different points (each with probability “p”) and result in effective response rates π , π , π , and π . Only results of setting 1 are shown in the main text, the rest in the supplements
| Setting | πSoC |
|
|
| Description |
|---|---|---|---|---|---|
| 1 | 0.10 | 0.20 (2) |
0.10 (1) with p = 0.5 0.20 (2) with p = 0.5 |
0.20 (1) if 0.40 (1) if | Backbone monotherapy superior to SoC, add‐on monotherapy has 50:50 chance to be superior to SoC; in case add‐on monotherapy not superior to SoC, combination therapy as effective as backbone monotherapy, otherwise combination therapy significantly better than monotherapies |
| 2 | 0.10 | 0.20 (2) | 0.10 (1) | 0.20 (1) | Backbone monotherapy superior to SoC, but add‐on monotherapy not superior to SoC and combination therapy not better than backbone monotherapy |
| 3 | 0.10 | 0.20 (2) | 0.10 (1) | 0.30 (1.5) | backbone monotherapy superior to SoC and combination therapy superior to backbone monotherapy, but add‐on monotherapy not superior to SoC |
| 4 | 0.10 | 0.20 (2) | 0.10 (1) | 0.40 (2) | Backbone monotherapy superior to SoC and combination therapy superior to backbone monotherapy (increased combination treatment effect compared to setting 4), but add‐on monotherapy not superior to SoC |
| 5 | 0.10 | 0.20 (2) | 0.20 (2) | 0.20 (0.5) | Both monotherapies are superior to SoC, but combination therapy is not better than monotherapies |
| 6 | 0.10 | 0.20 (2) | 0.20 (2) | 0.30 (0.75) | Both monotherapies are superior to SoC and combination therapy is better than monotherapies |
| 7 | 0.10 | 0.20 (2) | 0.20 (2) | 0.40 (1) | Both monotherapies are superior to SoC and combination therapy is superior to monotherapies (increased combination treatment effect compared to setting 7) |
| 8 | 0.10 | 0.10 (1) | 0.10 (1) | 0.10 (1) | Global null hypothesis |
| 9 | 0.20 | 0.20 (1) | 0.20 (1) | 0.20 (1) | Global null hypothesis with higher response rates |
| 10 | 0.10 | 0.20 (2) |
0.10 (1) with p = 0.5 0.20 (2) with p = 0.5 |
0.20* 0.20* 0.20* | Backbone monotherapy superior to SoC, add‐on monotherapy has 50:50 chance to be superior to SoC; combination therapy interaction effect can either be antagonistic/non‐existent, additive or synergistic (with equal probabilities) |
| 11 | 0.10 + 0.03*(c‐1) | 0.10 + 0.03*(c‐1) (1) | 0.10 + 0.03*(c‐1) (1) | 0.10 + 0.03*(c‐1) (1) | Time‐trend null scenario; every new cohort (first cohort |
| 12 | 0.10 + 0.03*(c‐1) | 0.20 + 0.03*(c‐1) (2) | 0.20 + 0.03*(c‐1) (2) | 0.40 + 0.03*(c‐1) (1) | Time‐trend scenario, whereby monotherapies superior to SoC and combination therapy superior to monotherapies; every new cohort (first cohort |
| 13 | 0.20 | 0.30 (1.5) | 0.30 (1.5) | 0.40 ( | Analogous to setting 7, but SoC response rate is 20% |
| 14 | 0.20 | 0.30 (1.5) | 0.30 (1.5) | 0.50 ( | Analogous to setting 8, but SoC response rate is 20% |
FIGURE 3Impact of the data sharing (linetype and point shape) and maximum number of cohorts per platform (x‐axis) on the per‐cohort and per‐platform type 1 error (Figure 3A) and power (Figure 3B) in treatment efficacy setting 1. Please note that different scaling of the y‐axis is used in the two subfigures. With increasing number of cohorts in the platform, the chance to make at least one correct positive or negative decision increases. When data is shared, the per‐cohort power increases, while it stays constant when no data sharing is planned
FIGURE 4Per‐cohort and per‐platform power with respect to data sharing (linetype and point shape), assumptions regarding the maximum number of cohorts (rows), cohort inclusion rate (columns) and final cohort sample size (x‐axis) in treatment efficacy setting 1. Generally, both types of power increase with increasing final cohort sample size. In case of no data sharing, the per‐cohort power is independent of assumptions regarding the expected platform size and cohort inclusion rate
FIGURE 5Impact of SoC response rate (columns; treatment efficacy settings 7 vs. 14), final cohort sample size (x‐axis) and data sharing (linetypes and point shapes) on per‐cohort and per‐platform power. We observed consistently lower power for increased SoC response rate. We also observed that while for lower sample sizes the per‐platform power is increased with increasing amount of data sharing, this is not true anymore for larger sample sizes
FIGURE 6Impact of the Bayesian GO decision rules on (A) per‐cohort and per‐platform type 1 error rates and (B) per‐cohort and per‐platform power in treatment efficacy setting 1. The black dot corresponds to the decision rules chosen in section 2.2. For every error rate, we set the data sharing (rows) to either full (“all”) or none (“cohort”). The x‐axis shows the required confidence (“gamma”) and the y‐axis the required superiority (“delta”) used in the Bayesian decision making in section 2.2. It is apparent that by choice of δ and γ, a wide range of different error rates can be obtained