| Literature DB >> 32440224 |
Kristian Thorlund1,2, Louis Dron2, Jay J H Park2,3, Edward J Mills1,2.
Abstract
There has been a rapid expansion in the use of non-randomized evidence in the regulatory approval of treatments globally. An emerging set of methodologies have been utilized to provide greater insight into external control data used for these purposes, collectively known as synthetic control methods. Through this paper, we provide the reader with a set of key questions to help assess the quality of literature publications utilizing synthetic control methodologies. Common challenges and real-life examples of synthetic controls are provided throughout, alongside a critical appraisal framework with which to assess future publications.Entities:
Keywords: RCTs; real-world evidence; synthetic control
Year: 2020 PMID: 32440224 PMCID: PMC7218288 DOI: 10.2147/CLEP.S242097
Source DB: PubMed Journal: Clin Epidemiol ISSN: 1179-1349 Impact factor: 4.790
Synthetic Control Quality Checklist
| Item Number | Key Question | Criteria for Judgement |
|---|---|---|
| External Control Data Sources | ||
| 1 | Was the original data collection process similar to that of the clinical trial? | State whether patients are from large well-conducted RCT(s) or high-quality prospective cohort studies, and whether patient characteristics are similar to the target population |
| 2 | Was the external control population sufficiently similar to the clinical trial population? | State how the external population is similar with regards to key characteristics, such as (but not limited to): age, geographic distribution, performance status, treatment history, sex etc. |
| 3 | Did the outcome definitions of the external control match those of that clinical trial? | State whether the outcomes are measured similarly or not |
| 4 | Was the synthetic control data set sufficiently reliable and comprehensive? | State whether there is sufficient sample sizes and covariates that can create comparable control groups |
| 5 | Were there any other major limitations to the dataset? | State any other potential limitations of the dataset that would limit the reliability and validity of comparisons |
| Synthetic Control Methods | ||
| 6 | Did the clinical trial include a concurrent control arm, or is the synthetic control data the only control data? | State the size of the concurrent control arm and whether the external data set is the only dataset being used or is being used to complement concurrent control arm(s) |
| 7 | How was the synthetic control data matched to the intervention group? | State the analytical method(s) – eg propensity matching scores – used to create the synthetic control arm |
| 8 | Were the results robust to sensitivity assumptions and potential biases? | State whether the sensitivity analyses were undertaken or reasons for not conducting sensitivity analyses, and compare whether the sensitivity analyses were comparable to the primary analyses. |
| 9 | Were synthetic control comparisons possible for all clinically important outcomes? | State if all clinically important outcomes were considered for analyses. If not, state justifications for not including all important outcomes |
| 10 | Are the results applicable to your patients? | State whether the synthetic control group created are similar to the patient group of interest |
| 11 | Were there any other major limitations to the synthetic control methods? | State any other potential limitations of the statistical methods that would limit the reliability and validity of comparisons |
Figure 1Quality check process.
A Summary of Commonly Used Models and Methods for Generating Synthetic Control Arms
| Model Complexity | Examples | Pros | Cons |
|---|---|---|---|
| Naïve | Simple mean, median or fixed-effect pooling | Easy to perform. | Requires high congruence between external and internal data. |
| Imbalance Adjustments | Multivariate regression, propensity scoring | Adjusts for imbalance to the extent explanatory factors are available in data. | Methods can be complex or relatively time consuming to implement and test. |
| Complex adjustment and weighting | Bayesian mixed-model commensurate power priors. | Can restore patient balance and weigh the contribution of multiple sources of data adequately. | Difficult and complex to implement. |
| Advanced exploratory solutions | Random forests, Neural Networks, Cluster analysis (Gaussian mixture models) | Can identify homogeneous sources of data for enhanced validity. | Mostly exploratory in nature and requires separate statistical analysis to produce synthetic control. |
Figure 2Schematic representation of dynamic borrowing or propensity-based methods. Adjustment methodologies refer to techniques such as propensity weighting or Bayesian dynamic borrowing as described in greater detail within Table 2. Here, an external dataset is adjusted utilizing statistical methodologies to make it more representative of a target population. This can either be centered on a target population existing as a partial control in a clinical trial, or a target treatment population. Depending on the method, a variable proportion of data from the external source is borrowed, adjusted for statistical (dis)similarity.
Figure 3Matching microsimulations are shown as utilizing external individual-level patient data to construct simulated cohorts which can represent truly synthetic control groups at the individual-patient level. Here, external data informs patient trajectories for the outcome(s) of interest to the relevant trial population, which can then be analyzed and compared to data from an interventional treatment arm.