| Literature DB >> 34386470 |
Jean C Digitale1, Kristefer Stojanovski2, Charles E McCulloch1, Margaret A Handley1,3,4.
Abstract
Background: In the face of the novel virus SARS-CoV-2, scientists and the public are eager for evidence about what measures are effective at slowing its spread and preventing morbidity and mortality. Other than mathematical modeling, studies thus far evaluating public health and behavioral interventions at scale have largely been observational and ecologic, focusing on aggregate summaries. Conclusions from these studies are susceptible to bias from threats to validity such as unmeasured confounding, concurrent policy changes, and trends over time. We offer recommendations on how to strengthen frequently applied study designs which have been used to understand the impact of interventions to reduce the spread of COVID-19, and suggest implementation-focused, pragmatic designs that, moving forward, could be used to build a robust evidence base for public health practice.Entities:
Keywords: COVID-19; difference-in-differences; implementation science; interrupted time series; preference design; sequential multiple assignment randomized trial; stepped wedge; study design
Mesh:
Year: 2021 PMID: 34386470 PMCID: PMC8353119 DOI: 10.3389/fpubh.2021.657976
Source DB: PubMed Journal: Front Public Health ISSN: 2296-2565
Overview of quasi-experimental designs.
|
|
|
|
|
|
|---|---|---|---|---|
| Pre-post | •Comparison of outcome of interest before and after intervention. | •Less cumbersome and simpler to gather data for than other designs (requires data from a minimum of only 2 time points). | •Temporal biases are a key threat to validity; if there are changes in measurement or quality of data over time, this will cause bias. | •Include comparator groups. |
| Interrupted time series (without control group) | •Data collected at multiple time points before and after an intervention is implemented. | •Each group acts as its own control. | •Requires a large number of measurements. | •Include comparator groups |
| Interrupted time series (with control group) | •Data collected at multiple time points before and after an intervention is implemented in a treatment group and control group. | •Controls for observed and unobserved time-invariant variables that differ between groups. | •Requires a large number of measurements. | •Evaluate parallel trends assumption. |
Selected examples of quasi-experimental studies evaluating real-world interventions to prevent COVID-19.
|
|
|
|
|
|
|---|---|---|---|---|
|
| ||||
| School closures (Mandate by the Qatari government) | •Compared rate of positive tests for respiratory viruses other than SARS-CoV-2 in a pediatric emergency department before and after school closures in Qatar ( | •Specified lag period for influenza A. | •No control group. | •Comparison group would improve validity. |
|
| ||||
| Physical distancing (Closures of schools, workplaces, and public transport, restrictions on mass gatherings/public events, and restrictions on movements [lockdowns]) | •Assessed incidence of COVID-19 before and after implementation of physical distancing interventions in 149 countries or regions, synthesized using meta-analysis ( | •Compared effect of five physical distancing interventions overall and in smaller subsets of policies to attempt to determine the most effective combination and sequence. | •No control group that was not subjected to at least one intervention. | •Comparison of “similar” clusters of countries (i.e., East African nations, Scandinavian nations) could improve analyses & interpretation. |
| Mask mandate (Universal mask wearing required by health system for healthcare workers and patients) | •Compared SARS-CoV-2 infection rate among healthcare workers before and after implementing universal masking in one health care system in the US ( | •Allowed for non-linear functional form of SARS-CoV-2 positivity rate. | •Testing was implemented for healthcare workers, but didn't fully account for lags in development of symptoms after implementation of policy in their division of time. | •Add comparison group. |
| Social distancing measures (closures of schools, closures of workplaces, cancellations of public events, restrictions on internal movement, and closures of state borders) | •Estimated change in COVID-19 case growth and mortality before and after implementation of first statewide social distancing measures ( | •Specified an event-study design as a robustness check. | •The type of the first social distancing measure may have differed across states. | •Exploration of how lifting of policies, as compared to those who kept policies (i.e., duration of intervention), could improve interpretation. |
| School closures (State government mandates) | •Assessed whether school closures impacted incidence of COVID-19 at the beginning of the pandemic in the US ( | •Included other non-school related policies (e.g., stay at home orders) in models. | •No control group. | •Localized nature of policies could provide advantage for cluster ITS comparisons, as compared to state-level data used in the study. |
|
| ||||
| Stay-at-home orders (State government mandates) | •Compared COVID-19 cases in border counties in Illinois (where a stay-at-home order was issued) to border counties in Iowa (where such an order was not issued) ( | •Comparison of border counties potentially less likely to be biased than comparison of larger geographic area. | •Only one pre-period, as compared to six post-periods. | •Inclusion of analyses of sequencing of orders in Iowa could strengthen analysis. |
| Social distancing measures (Bans on large social gatherings; school closures; closures of entertainment venues, gyms, bars, and restaurant dining areas; and shelter-in-place orders) | •Assessed effect of social distancing measures on measures of growth rate of confirmed COVID-19 cases in US counties using an event study design ( | •Event study design (including fixed effects for county and time) allowed testing of parallel trends assumption in pre-policy period. | •Relying on administrative boundaries such as counties may not reflect how people live their lives (e.g. working across county lines), making it more difficult to interpret findings. | •Could have used localized data to make comparisons over time, comparing similar states (clusters) with more or less restrictive orders. This is particularly important given that controlling for number of tests was done at the state-level, not locally. |
| •Inclusion of a longer pre-intervention period would improve the study; could have used excess mortality as a marker of COVID-19 cases. | ||||
| Face mask mandates (State government policies to wear face masks or covers in public) | •Assessed effect of state government mandates for face mask use on changes in daily US county-level COVID-19 growth rates using an event study design ( | •Event study design allowed testing of parallel trends assumption in pre-policy period. | •Some states did not have state-wide mandates, but counties within them enacted mandates. | •Local-level variation in adherence to mandates could alter results, comparison of county adherence measures (e.g., fines) could strengthen analyses. |
Figure 1Interrupted time series. An example of an interrupted time series design with no control group. A scatterplot of data is shown with the intervention implemented at the time of the dotted line. This plot indicates a level change (but no slope change) due to the intervention.
Figure 2Interrupted time series with control group. An example of an interrupted time series design with control group (often analyzed with a difference-in-differences approach). A scatterplot of data is shown with an intervention (orange) and control (green) group. The intervention is implemented in the treated group at the time of the vertical dotted line. The orange dashed line refers to the hypothetical outcome of the treated group in the absence of the intervention. The difference between this hypothetical outcome and the actual outcome is the treatment effect.
Pragmatic study design examples applied to community interventions during COVID-19.
|
|
|
|
|
|---|---|---|---|
| Two-stage randomized preference | Health departments would like to offer incentives to contacts to remain in quarantine for the full recommended duration but cannot offer all types of possible incentives so would like to determine which are more impactful. | •Participation/engagement levels for those randomized to different options vs. randomization to preference. | |
| SMART | Clinic systems may not be able to offer video-visits to all patients, and can benefit from determining whether less intensive formats (e.g., telephone calls; email communications) are sufficient for some patients, allowing the more intensive formats to be offered to those who struggle with other formats, or whose health needs do not align with less intensive formats. | Comparisons of: (1) patient-level and provider-level engagement with different telemedicine options; (2) levels of satisfaction; (3) outcome metrics such as completion of referrals, labs, refills of patients in different groups/no-show rates at the clinic. | |
| Stepped wedge design (modified) | Schools may want to re-open but prefer a staggered approach, in which all schools start with on-line learning, and then depending on outcomes of COVID-19 testing after the school starts, changes in restrictions are made, such as in-person attendance. | •Do the schools/classrooms meet the advancement criteria for moving to the next school reopening level? |
Figure 3Two-Stage Preference Design for Contact Tracing Quarantine Incentives. The comparison of uptake of A1 vs. B1 shows the selection effect. Is there differential uptake of these two programs? If yes, then there is a difference in the groups' overall selection likelihood. The comparison of outcomes of A2 vs. B2 shows the difference between two programs through a controlled trial design. For example, for the research question: Is there a difference in measures of successful completion of quarantine between the two programs? The comparison of outcomes of A1 vs. A2 and B1 vs. B2 shows the preference effect. For example, if more participants who selected cash stipend (A1), were likely to complete their second COVID-19 test than those who were randomized to cash stipend (A2).
Figure 4SMART Design for Telemedicine Visit Type in Primary Care. Individuals are initially randomized (R in circle) to either telephone visits or video visits. Those who are not responding to the intervention are re-randomized to continue the same intervention, switch interventions, or add a health coach call.
Figure 5SMART/Stepped Wedge Design for School Re-Opening. Credit: Dr. Naomi Bardach, University of California, San Francisco. In this design, Steps 1–3 each represent an increasing number of in-person students. The team will conduct baseline: (1) PCR COVID-19 testing at all schools, for students and teachers and staff, and (2) student and teacher surveys regarding exposure and symptom history. Then, weekly PCR testing for a random sampling of students and staff within each school cluster will be conducted to determine if changes from Step 1 to Step 2 will be allowable after 3 weeks. If no new outbreaks occur during the move to Step 2, nor during the weeks 9–11 when all schools are in Step 2, all school clusters will be newly randomized and move to Step 3 practices. If no or limited outbreaks occur, we will recommend staying in Step 3 restrictions. Should there be large outbreaks or several small outbreaks in any of the schools in any of the stages, schools can return to the more restrictive Step 2 practices.