| Literature DB >> 32046735 |
Michael W Beets1, R Glenn Weaver2, John P A Ioannidis3, Marco Geraci2, Keith Brazendale2, Lindsay Decker2, Anthony D Okely4, David Lubans5, Esther van Sluijs6, Russell Jago7, Gabrielle Turner-McGrievy2, James Thrasher2, Xiaming Li2, Andrew J Milat8,9.
Abstract
BACKGROUND: Preliminary evaluations of behavioral interventions, referred to as pilot studies, predate the conduct of many large-scale efficacy/effectiveness trial. The ability of a pilot study to inform an efficacy/effectiveness trial relies on careful considerations in the design, delivery, and interpretation of the pilot results to avoid exaggerated early discoveries that may lead to subsequent failed efficacy/effectiveness trials. "Risk of generalizability biases (RGB)" in pilot studies may reduce the probability of replicating results in a larger efficacy/effectiveness trial. We aimed to generate an operational list of potential RGBs and to evaluate their impact in pairs of published pilot studies and larger, more well-powered trial on the topic of childhood obesity.Entities:
Keywords: Childhood obesity; Diet; Framework; Intervention; Physical activity; Scalability; Screen time; Sleep; Youth
Mesh:
Year: 2020 PMID: 32046735 PMCID: PMC7014944 DOI: 10.1186/s12966-020-0918-y
Source DB: PubMed Journal: Int J Behav Nutr Phys Act ISSN: 1479-5868 Impact factor: 6.457
Examples of Generalizability Biases in the Childhood Obesity Literature
| Bias | Likely Larger Effect | Likely Smaller/No Effect |
|---|---|---|
| Study | Fitzgibbon 2005 [ | Kong 2016 [ |
| Who delivered the intervention? | “…the use of specially trained early childhood educators rather than classroom teachers to deliver the intervention, thereby raising questions of generalizability.” | “…using teachers in existing Head Start classrooms to deliver the intervention.” |
| Study | Cohen 2015 [ | Sutherland 2017 [ |
| How much of the intervention was provided? | 1 full day training and 1 half day training | 1 90-min training |
| Study | Beets 2016 [ | Beets 2018 [ |
| How much support to implement the intervention was provided? | “During the first year of receiving the intervention for both the immediate and delayed program, each program received four booster sessions. During the second year of receiving the intervention (for the immediate condition only) 2 booster sessions/program were provided.” | No additional onsite booster sessions or follow-up |
| Study | Sutherland 2016 [ | |
| Who delivered the intervention? | “The provision of an in-school physical activity consultant for 1 day per week was the largest cost relating to the efficacy trial (66% of the total intervention cost). Whilst the provision of an in-school physical activity consultant was necessary under efficacy trial conditions in order to evaluate the effect of the combination of intervention strategies, the feasibility of providing a part-time consultant within schools across large geographic regions and the cost of such a model of support presents challenges in upscaling the intervention. The dissemination of an effective intervention across the community requires the use of implementation strategies which better mirror real world practice.” | |
| Study | McKenzie 1996 [ | Hoelscher 2004 [ |
| How much support to implement the intervention was provided? | “Following initial training, CATCH PE consultants provided on-site follow-up approximately every 2 weeks. During the 2.5 years, consultants made 3089 documented school visits, averaging 55.3 per school and 51.7 min in length. Consultants performed various roles during visits, including giving feedback to teachers, modeling new lesson segments, team teaching, and providing motivation and technical support.” | No onsite, on-going support provided |
| Study | Salmon 2008 [ | Salmon 2011 [ |
| How much of the intervention was provided? | 19 lessons delivered | 6 lessons delivered “…Switch-2-Activity involved an abbreviated programme; therefore, the intervention ‘dose’ was lower…” |
| How long was the intervention delivered? | 10 months | 7 weeks |
| Who delivered the intervention? | “All intervention components were delivered by one intervention specialist (a qualified Physical Education teacher) across all three schools.” | “the programme was delivered by regular class teachers rather than by a specialist university research team…” |
| What measures were used to collect information on outcomes? | Objective measures | Self-report |
| Study | West 2010 [ | Gerards 2015 [ |
| Who delivered the intervention? | “All sessions were facilitated by a clinical psychologist and accredited provider of the intervention (who co-authored the intervention materials), with assistance from graduate students in nutrition and dietetics, physical education, and psychology.” | “The intervention was led by three different facilitators. These health professionals have been accredited after attending an official 3-day training course and an additional intervention day.” “Finally, the West 2010 [ |
| Who received the intervention? | “participants were mainly white, well-educated parents with moderate levels of employment and income.” |
Operational Definitions of Risk of Generalizability Biases
| Risk of Generalizability Bias | Questions to Ask | Increased Presence with Small Sample | Hypothesized Influence of the Presence of Risk of Generalizability Bias | Example | ||
|---|---|---|---|---|---|---|
| Pilot | Larger-Scale Efficacy/Effectiveness | Pilot | Larger-Scale Efficacy/Effectiveness | |||
| What is the potential for difference(s) between… | ||||||
| Intervention Intensity Bias | …the number and length of contacts in the current study and future evaluations of the intervention? | Yes | More frequent and longer contacts result in more effective intervention | Fewer and shorter contacts results in less effective intervention compared to pilot | 19 lessons delivered (Salmon 2008 [ | 6 lessons delivered (Salmon 2011 [ |
| Implementation Support Bias | …the amount of support provided to implement the intervention in the current study and future evaluations of the intervention? | Yes | Greater amounts of support to implement the intervention results in more effective intervention | Reduced support to implement the intervention results in less effective intervention compared to pilot | ||
| Intervention Delivery Agent Bias | …the level of expertise of the individual(s) who deliver the intervention in the current study compared to who will deliver the intervention in future evaluations? | Yes | Higher levels of expertise delivering the intervention results in more effective intervention | Lower level of expertise to deliver the intervention results in less effective intervention compared to pilot | ||
| Target Audience Bias | …the demographics of those that received the intervention in the current study to those who will receive the intervention in future evaluations? | No | Delivering intervention to more conducive, convenience sample or sample that is not representative of target population results in more effective intervention | Delivering intervention to sample of whom the intervention is intended results in less effective intervention compared to pilot | ||
| Intervention Duration Bias | …the length of the intervention provided in the current study to the length of the intervention in future evaluations? | No | Shorter duration results in more effective intervention | Longer duration less effective intervention compared to pilot | 4-week intervention (Wilson 2005 [ | 17-week intervention (Wilson 2011 [ |
| Setting Bias | …the setting where the intervention is delivered in the current study and the intervention delivery setting in future evaluations? | No | Delivering intervention in a more conducive, convenience location that is not representative of the target setting results in more effective intervention | Delivering intervention in a location more representative of target setting results in a less effective intervention compared to pilot | Intervention delivered on university campus b | Intervention delivered in community setting b |
| Measurement Bias | …the measures employed in the current study and the measures used in future evaluations of the intervention for primary/secondary outcomes? | Yes | Use of less reliable or valid measures of primary/secondary outcomes results in more effective intervention | Use of more reliable and valid measures results in less effective intervention compared to pilot | Pedometer used to measure physical activity (Lubans 2009 [ | Accelerometer used to measure physical activity (Lubans 2012 [ |
| Directional Conclusions | Are the intervention effect(s) in the hypothesized direction? | No | Less effective intervention | Reduces intervention effectiveness | ||
| Outcome Bias | Is the primary outcome for future evaluations of the intervention measured in the current study? | No | Absences of measuring primary outcome results in more effective intervention | Absence of primary outcome collected in pilot results in less effective intervention tested in well-powered trial | Nutrients sold per day and number of items sold per day in school cafeterias (Hartstein 2008 [ | Self-reported daily dietary intake of students (Siega-Riz 2011 [ |
aAlthough not labeled as a pilot study, the example illustrates the presence of the risk of generalizability bias in one study and altered in the subsequent trial
bHypothetical example of the risk of generalizability bias as it could operate in a pilot to larger-scale efficacy/effectiveness trial
Fig. 1PRISMA diagram of literature search
Fig. 2Presence of risk of generalizability biases in pilot and larger-scale efficacy/effectiveness pairs. Note: Red circle () indicates bias present, green circle () bias not present, orange circle () bias identified in pilot or well-powered but not the other. E-E = Efficacy/Effectiveness. a Sample size represents setting level (e.g., school, childcare) – child-level sample size not reported
Fig. 3Forest plot of the change in the standardized mean difference (SMD) of the presence, absence, or carry forward of six risk of generalizability biases from a pilot to larger-scale efficacy/effectiveness (E/E) trial
Fig. 4Association of the three most prevalent risk of generalizability biases with pilot and efficacy/effectiveness sample size. Note: The x- and y-axis represent the log of the total sample size per study. The tick marks represent the actual total sample size across the range of sample sizes in the studies.