Literature DB >> 31819600

Number Of Clinical Trial Study Sites Impacts Observed Treatment Effect Size: An Analysis Of Randomized Controlled Trials Of Opioids For Chronic Pain.

Diana S Meske1, Ben J Vaughn2, Ernest A Kopecky3, Nathaniel Katz4,5.   

Abstract

BACKGROUND: Many aspects of study conduct impact the observed effect size of treatment. Data were utilized from a recently published meta-analysis of randomized, double-blind, placebo-controlled, clinical trials performed for the United States Food and Drug Administration (FDA) approval of full mu-agonist opioids for the treatment of chronic pain.
METHODS: The number of study sites in each clinical trial and standardized effect size (SES) were extracted and computed. Standardized effect size was plotted against number of sites, and a two-piece linear model was fit to the plot. Ten studies were included.
RESULTS: The SES decreased linearly by 0.13 units for every 10 sites (p=0.037), from 0.75 to 0.36, until an inflection point of 60 sites, after which SES did not decline further. The total number of subjects required for 90% power to discriminate drug from placebo increased from 78 to 336 subjects going from 30 to 60 sites.
CONCLUSION: Results showed that the number of sites was a source of loss of assay sensitivity in clinical trials, which may contribute to the well-known problem of failure to successfully transition from Phase 2 to Phase 3 clinical development. Potential solutions include minimizing the number of sites, more rigorous and validated training, central statistical monitoring with rapid correction of performance issues, and more rigorous subject and site selection.
© 2019 Meske et al.

Entities:  

Keywords:  chronic pain; clinical trials; effect size; opioids; randomized controlled trials; study site number

Year:  2019        PMID: 31819600      PMCID: PMC6875508          DOI: 10.2147/JPR.S201751

Source DB:  PubMed          Journal:  J Pain Res        ISSN: 1178-7090            Impact factor:   3.133


Introduction

The United States Food and Drug Administration (FDA) requires randomized, double-blind, placebo-controlled studies of at least a 3-month duration to approve treatments for chronic diseases such as pain.1 These studies typically take years to enroll and pose a significant financial burden on companies and institutions seeking regulatory approval. To expedite enrollment, many sponsors are tempted to utilize large numbers of study sites. However, inconsistencies in study conduct at each study site have long been assumed to add variability to the data2 and may increase the likelihood of a failed study (i.e., the treatment actually is efficacious, but the study fails).3 Results of small, carefully performed, successful Phase 2 studies have long been observed to be difficult to replicate in large, multi-site, Phase 3 studies, even when sample size is increased.4 There are two possible explanations: 1) the smaller Phase 2 study results were falsely positive or 2) due to larger error variability, the Phase 3 study was falsely negative, or lacked adequate assay sensitivity. Opioids are a useful class of drugs for discerning the impact of study design and conduct on observed effect sizes because, in principle, all full mu-opioid agonists have the same pharmacologic mechanism, binding to the mu-opioid receptor, and therefore produce the same therapeutic effect once doses are optimized.5 A 2005 qualitative review of all randomized controlled trials (RCTs) of opioids found that the fewer the sites, the larger the observed standardized effect size (SES) of treatment.6 The relationship between numbers of sites and observed treatment effects has seldom been quantitatively evaluated. This study reanalyzes data previously collected for a meta-analysis of clinical trials of opioids versus placebo for chronic pain submitted for regulatory approval in the US,7 to characterize the relationship between number of sites and SES, and to suggest an appropriate adjustment to SES for sample size calculations based on the number of study sites the sponsor intends to use in a Phase 3 clinical trial.

Materials And Methods

This analysis utilized data collected from a recent meta-analysis of opioid studies for chronic pain.7 A detailed description of the methods of the original study can be found in the original publication. In this study, all published double-blind, randomized, placebo-controlled, enriched enrollment randomized withdrawal (EERW) studies that compared any full mu-agonist opioid (excluding partial mu-agonists and mixed activity opioids) to placebo for ≥12 weeks during the randomized, double-blinded treatment phase were selected. Any type of chronic non-cancer pain was allowed; acute or cancer pain was excluded. Agents with oral, transdermal, nasal, sublingual, and transmucosal routes of administration were included. The original study protocol was published on PROSPERO (registration #: CRD42015026378). The primary endpoint of the current analysis was the relationship across studies between the number of study sites and SES. The number of study sites in which each clinical trial was conducted was extracted, as were sufficient data to compute an SES (change in pain intensity [PI] from randomization baseline to week 12 or study endpoint and its variability); when necessary, this information was extracted from other publicly available sources (e.g., FDA Summary Basis of Approval, ). Standardized effect size for each study was defined as [{(change from baseline to week 12 in PI on drug) – (change from baseline to week 12 in PI on placebo)}/(pooled standard deviation of the change from baseline in PI)]. SES is independent of the study’s sample size, but studies with larger sample sizes will have more precise estimates of the SES. Observed SES was plotted against the number of sites utilized for each protocol and a smoothed line was fit to the plot to examine the overall shape. For sensitivity, this was jackknifed by removing one protocol at a time from the sample and repeating the analysis. This revealed a consistent inflection point at around 60 sites followed by a flat trend line beyond 60 sites. This was then used for selection of the knot point in the final two-piece linear model. The number of subjects that theoretically would be required for 90% power to detect the observed SES at participating sites was then calculated based on the model-predicted SES for a given site count and plotted.

Results

Fifteen EERW studies were included in the original meta-analysis. Five were excluded because they studied drugs were not full mu-agonists or had additional mechanisms of action (monoamine reuptake inhibition), namely buprenorphine, tramadol, and tapentadol, leaving 10 studies (see Table 1). Drugs evaluated were hydrocodone (4 clinical trials); oxymorphone and oxycodone (2 clinical trials each); and hydromorphone and morphine/naltrexone (1 clinical trial each). Disorders studied included chronic low back pain and osteoarthritis. Standardized effect sizes were based on the group mean difference between drug and placebo at the end of the randomized, double-blind treatment period, and ranged from 0.173 to 0.913. The estimated mean SES of the 10 studies included in this analysis as assessed by binary random-effects model meta-analysis using the restricted maximum likelihood method is 0.399 (95% CI: 0.239, 0.559). Further details are available in the original publication.7
Table 1

Enriched Enrollment, Randomized Withdrawal Studies Of Full Mu Opioid Agonists For The Treatment Of Chronic Pain Submitted For FDA Approval Included In This Analysis

AuthorYearDrug, Pain TypeN# Of Sites, CountriesChange In PI-NRS ScoreMean (SD)SESEstimate (95% CI)
Hale8 2007Oxymorphone ER, CLBP14230, USDrug:8.7 (25.1)0.913 (0.567, 1.259)
Placebo:31.6 (24.6)
Katz9 2007Oxymorphone ER, CLBP20529, USDrug:10.9 (24.5)0.574 (0.294, 0.853)
Placebo:26.0 (27.9)
Hale10 2010Hydromorphone ER, CLBP26666, USDrug:0.6 (1.3)0.799 (0.550, 1.049)
Placebo:1.7 (1.5)
Katz11 2010Morphine – sequestered naltrexone ER capsules, OA34374 USDrug:−0.2 (1.9)0.249 (0.037, 0.462)
Placebo:0.3 (2.1)
Friedman12 2011 Oxycodone ER, OA41061, USDrug:−0.7 (2.1)0.173 (−0.021, 0.367)
Placebo:−0.3 (2.5)
Rauck13 2014Hydrocodone ER, CLBP30259, USDrug:0.48 (1.6)0.308 (0.081, 0.535)
Placebo:0.96 (1.6)
Wen14 2015Hydrocodone bitartrate ER, CLBP588102, USDrug:3.7 (1.9)0.267 (0.104, 0.429)
Placebo:4.2 (2.0)
Katz15 2015Oxycodone ER, CLBP38946, USDrug:0.29 (2.1)0.592 (0.389, 0.795)
Placebo:1.85 (3.1)
Hale16 2015aHydrocodone ER, CLBP/OA29371, USDrug:−0.64 (1.6)0.347 (0.117, 0.578)
Placebo:−0.03 (1.9)
Hale17 2015bHydrocodone ER, CLBP37178, USDrug:0.11 (1.9)0.319 (0.114, 0.524)
Placebo:0.74 (2.0)

Note: Data from Meske et al.7

Abbreviations: CLBP, chronic low back pain; CI, confidence interval; ER, extended-release; N, number of subjects included in the randomized, double-blind, placebo-controlled portion of the study; OA, osteoarthritis; PI-NRS, pain intensity – numerical rating scale; SD, standard deviation; SES, standardized effect size; US, United States.

Enriched Enrollment, Randomized Withdrawal Studies Of Full Mu Opioid Agonists For The Treatment Of Chronic Pain Submitted For FDA Approval Included In This Analysis Note: Data from Meske et al.7 Abbreviations: CLBP, chronic low back pain; CI, confidence interval; ER, extended-release; N, number of subjects included in the randomized, double-blind, placebo-controlled portion of the study; OA, osteoarthritis; PI-NRS, pain intensity – numerical rating scale; SD, standard deviation; SES, standardized effect size; US, United States. The relationship between number of sites and observed SES is presented graphically in Figure 1, which also illustrates the results of the two-piece linear model (p-value=0.037). The slope of the model declined at a rate of 0.13 SES units per 10 sites from approximately 30 to 60 sites (i.e., for every additional 10 sites, the SES declined by 0.13 units), then was flat (i.e., adding additional sites was not associated with a further decline in SES). Because the smallest number of sites of any of the included studies was 29, there are no data below that number of sites.
Figure 1

A two-piece linear model of the relationship between number of study sites (x-axis) and observed SES of treatment (y-axis).

A two-piece linear model of the relationship between number of study sites (x-axis) and observed SES of treatment (y-axis). Figure 2 presents the total number of subjects required for 90% power to detect the observed difference under the model shown in Figure 1, contrasted with the number needed under the traditional assumption of no impact of number of sites on observed effect size. As clinical trial size increases from 30 to 60 sites, the total number of subjects required increases from 78 to 336. Beyond 60 sites, there is no further increase in sample size requirement because there was no further degradation of SES.
Figure 2

Total number of subjects required form 90% power, comparing the modeled actual relationship between number of sites and SES, vs the assumption of no impact of number of sites in SES.

Total number of subjects required form 90% power, comparing the modeled actual relationship between number of sites and SES, vs the assumption of no impact of number of sites in SES.

Discussion

The problem of variability in study conduct between clinical trial sites has been lamented for years in both scientific18–21 and regulatory2,22 documents. There appears to be only one previous observation on the relationship between number of sites and effect sizes, a meta-analysis of pediatric antidepressant clinical trials, which mentioned that the effect size was inversely proportional to the number of sites.23 To the authors’ knowledge, the current study is the first study to focus on the quantitative relationship between number of sites and effect sizes. Results showed that as the number of sites in clinical trials of opioids for chronic pain increases, the effect size decreases, until a plateau is reached where effect size does not decline further. The impact was large: a 30-site study, fully powered with 80 subjects, would need over 300 subjects if the number of sites were increased to 60. Thus, using data from Phase 2 studies conducted with fewer sites may seriously underestimate sample size requirements when applied to Phase 3 clinical trials with more sites. While there are good reasons to conduct multicenter clinical trials,19 utilizing many sites to accelerate timelines may lead to a “haste makes waste” phenomenon, where by racing to complete clinical trials as soon as possible, sponsors may end up at the starting line, as failed clinical trials need to be repeated. The ethical implications of experimenting on human subjects in inadvertently underpowered studies are also concerning. Moreover, some sponsors may not be able to withstand the financial consequences of a failed study; promising treatments may never make it to market. The measurement error that underlies the impact of site variability on assay sensitivity must be minimized to avoid false-negative clinical trials.3 Several studies have described methods to prevent measurement error by training patients to report their symptoms more accurately,24 screening out patients who are not able to report symptoms accurately,25,26 and a variety of other methods.3 Identification and remediation of measurement error during ongoing clinical trials by central statistical monitoring and intervention, as required by regulation,22 has also been described in a few studies.18,27 This analysis was limited by the number of available studies; the observed relationship could change with more studies. Only opioid clinical trials were included; clinical trials of other analgesics, or other therapeutic areas, may not show the same relationship. The two-slope function that explained this dataset may not be robust to other datasets. None of the included studies used fewer than 29 sites; findings cannot be extrapolated below this threshold. The smallest study included only 142 subjects, which may make it prone to bias; however, the remainder had greater than 200 subjects, making this less of a risk.28 In summary, this analysis of opioid clinical trials suggests that the greater the number of sites, the smaller the observed effect size of treatment. This means that more subjects overall, and per site, need to be added to preserve power as the number of sites increases. Sample size calculations from smaller Phase 2 studies may seriously underestimate sample size requirements for larger Phase 3 studies. This may be one of several causes of failed Phase 3 clinical trials; conclusions should be interpreted in the context of the limitations of this study.
  25 in total

1.  Pitfalls of multisite randomized clinical trials of efficacy and effectiveness.

Authors:  H C Kraemer
Journal:  Schizophr Bull       Date:  2000       Impact factor: 9.306

Review 2.  Methodological issues in clinical trials of opioids for chronic pain.

Authors:  Nathaniel Katz
Journal:  Neurology       Date:  2005-12-29       Impact factor: 9.910

3.  A 12-week, randomized, placebo-controlled trial assessing the safety and efficacy of oxymorphone extended release for opioid-naive patients with chronic low back pain.

Authors:  Nathaniel Katz; Richard Rauck; Harry Ahdieh; Tina Ma; Roland Gerritsen van der Hoop; Rosemary Kerwin; Gilbert Podolsky
Journal:  Curr Med Res Opin       Date:  2007-01       Impact factor: 2.580

Review 4.  Opioids: after thousands of years, still getting to know you.

Authors:  Nathaniel Katz
Journal:  Clin J Pain       Date:  2007-05       Impact factor: 3.442

5.  Considerations for improving assay sensitivity in chronic pain clinical trials: IMMPACT recommendations.

Authors:  Robert H Dworkin; Dennis C Turk; Sarah Peirce-Sandner; Laurie B Burke; John T Farrar; Ian Gilron; Mark P Jensen; Nathaniel P Katz; Srinivasa N Raja; Bob A Rappaport; Michael C Rowbotham; Misha-Miroslav Backonja; Ralf Baron; Nicholas Bellamy; Zubin Bhagwagar; Ann Costello; Penney Cowan; Weikai Christopher Fang; Sharon Hertz; Gary W Jay; Roderick Junor; Robert D Kerns; Rosemary Kerwin; Ernest A Kopecky; Dmitri Lissin; Richard Malamut; John D Markman; Michael P McDermott; Catherine Munera; Linda Porter; Christine Rauschkolb; Andrew S C Rice; Cristina Sampaio; Vladimir Skljarevski; Kenneth Sommerville; Brett R Stacey; Ilona Steigerwald; Jeffrey Tobias; Ann Marie Trentacosti; Ajay D Wasan; George A Wells; Jim Williams; James Witter; Dan Ziegler
Journal:  Pain       Date:  2012-04-09       Impact factor: 6.961

6.  Single-entity hydrocodone extended-release capsules in opioid-tolerant subjects with moderate-to-severe chronic low back pain: a randomized double-blind, placebo-controlled study.

Authors:  Richard L Rauck; Srinivas Nalamachu; James E Wild; George S Walker; Cynthia Y Robinson; Charles S Davis; Stephen J Farr
Journal:  Pain Med       Date:  2014-02-12       Impact factor: 3.750

7.  A phase 3, multicenter, randomized, double-blind, placebo-controlled, safety, tolerability, and efficacy study of Xtampza ER in patients with moderate-to-severe chronic low back pain.

Authors:  Nathaniel Katz; Ernest A Kopecky; Melinda O'Connor; Robert H Brown; Alison B Fleming
Journal:  Pain       Date:  2015-12       Impact factor: 6.961

8.  Clinical response and risk for reported suicidal ideation and suicide attempts in pediatric antidepressant treatment: a meta-analysis of randomized controlled trials.

Authors:  Jeffrey A Bridge; Satish Iyengar; Cheryl B Salary; Rémy P Barbe; Boris Birmaher; Harold Alan Pincus; Lulu Ren; David A Brent
Journal:  JAMA       Date:  2007-04-18       Impact factor: 56.272

9.  Ongoing monitoring of data clustering in multicenter studies.

Authors:  Lauren B Guthrie; Emily Oken; Jonathan A C Sterne; Matthew W Gillman; Rita Patel; Konstantin Vilchuck; Natalia Bogdanovich; Michael S Kramer; Richard M Martin
Journal:  BMC Med Res Methodol       Date:  2012-03-13       Impact factor: 4.615

10.  Efficacy and tolerability of a hydrocodone extended-release tablet formulated with abuse-deterrence technology for the treatment of moderate-to-severe chronic pain in patients with osteoarthritis or low back pain.

Authors:  Martin E Hale; Charles Laudadio; Ronghua Yang; Arvind Narayana; Richard Malamut
Journal:  J Pain Res       Date:  2015-09-15       Impact factor: 3.133

View more
  3 in total

Review 1.  Design and conduct of confirmatory chronic pain clinical trials.

Authors:  Nathaniel Katz
Journal:  Pain Rep       Date:  2020-12-18

2.  Research design considerations for randomized controlled trials of spinal cord stimulation for pain: Initiative on Methods, Measurement, and Pain Assessment in Clinical Trials/Institute of Neuromodulation/International Neuromodulation Society recommendations.

Authors:  Nathaniel Katz; Robert H Dworkin; Richard North; Simon Thomson; Sam Eldabe; Salim M Hayek; Brian H Kopell; John Markman; Ali Rezai; Rod S Taylor; Dennis C Turk; Eric Buchser; Howard Fields; Gregory Fiore; McKenzie Ferguson; Jennifer Gewandter; Chris Hilker; Roshini Jain; Angela Leitner; John Loeser; Ewan McNicol; Turo Nurmikko; Jane Shipley; Rahul Singh; Andrea Trescot; Robert van Dongen; Lalit Venkatesan
Journal:  Pain       Date:  2021-07-01       Impact factor: 6.961

Review 3.  Study Design Characteristics and Endpoints for Enriched Enrollment Randomized Withdrawal Trials for Chronic Pain Patients: A Systematic Review.

Authors:  David J Kopsky; Karolina M Szadek; Patrick Schober; Alexander F J E Vrancken; Monique A H Steegers
Journal:  J Pain Res       Date:  2022-02-17       Impact factor: 3.133

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.