| Literature DB >> 32748164 |
Nick Axford1, Vashti Berry2, Jenny Lloyd3, Tim Hobbs4, Katrina Wyatt3.
Abstract
There can be a tendency for investigators to disregard or explain away null or negative results in prevention science trials. Examples include not publicizing findings, conducting spurious subgroup analyses, or attributing the outcome post hoc to real or perceived weaknesses in trial design or intervention implementation. This is unhelpful for several reasons, not least that it skews the evidence base, contributes to research "waste", undermines respect for science, and stifles creativity in intervention development. In this paper, we identify possible policy and practice responses when interventions have null (ineffective) or negative (harmful) results, and argue that these are influenced by: the intervention itself (e.g., stage of gestation, perceived importance); trial design, conduct, and results (e.g., pattern of null/negative effects, internal and external validity); context (e.g., wider evidence base, state of policy); and individual perspectives and interests (e.g., stake in the intervention). We advance several strategies to promote more informative null or negative effect trials and enable learning from such results, focusing on changes to culture, process, intervention design, trial design, and environment.Entities:
Keywords: Evaluation; Negative effect; Null effect; Randomized controlled trial
Mesh:
Year: 2022 PMID: 32748164 PMCID: PMC7398716 DOI: 10.1007/s11121-020-01140-4
Source DB: PubMed Journal: Prev Sci ISSN: 1389-4986
Common researcher responses to finding null or negative effects in prevention science trials
| Response | Effect | Legitimacy |
|---|---|---|
| 1. | This contributes to a skewed impression of “what works” because the studies do not get picked up in systematic reviews and meta-analyses; specifically, evidence of effectiveness is likely to be exaggerated. | Failure to submit a results paper for publication is not necessarily a deliberate act, rather it can occur through inertia (although when an author is involved in intervention design or dissemination, this distinction becomes blurred). Journal editors and reviewers tend not to say that a lack of effect is the reason for rejection, but null effects rarely constitute the “ground-breaking” findings that journals invariably aspire to publish. |
| 2. | The chances of finding false-positive results from a single dataset increase as more hypotheses are tested, so this practice can produce misleading results. | Moderator analyses specified a priori in the trial protocol or statistical analysis plan can be suitable, even if they are exploratory and acknowledged to be underpowered. However, there is widespread agreement that it is inappropriate to conduct ad hoc or theoretically uninformed moderator analyses in an attempt to find a positive effect for a subgroup. |
| 3. | This creates the appearance of effectiveness, especially if the findings are “spun” in the write-up (e.g., by referring to “positive effects” in the abstract and relegating information about the lack of effect to the body text). | Given the difficulty of publishing null or negative findings, this response is unsurprising but problematic when it concerns any of the following: a secondary outcome or mediator; an interim data collection point; an outcome with marginal statistical significance (or the level of statistical acceptability is changed to make it “significant”); or a tiny effect that is unlikely to be of practical or clinical significance (even if it is |
| 4. | This casts doubt on the veracity of the findings (even when that is unfair), leading the reader to conclude that the intervention is potentially effective or of unknown effectiveness. | It is reasonable to identify limitations to trial methodology when reporting results, and for interested observers to critique the methods. Limitations in design or conduct might present a valid explanation for the lack of positive effects, with important implications for the interpretation of findings and conduct of future research. However, it is disingenuous to identify such problems only once results are known. |
| 5. | This suggests that the intervention would be effective if delivered as intended. | There is strong evidence for a positive association between fidelity and outcomes, so exploring this relationship is reasonable. However, care is needed not to use fidelity as an excuse once outcome results are known. Moreover, fidelity |
| 6. | This argument can be deployed to suggest that the intervention is effective but that it did not work | Contextual arguments may be legitimate, and can help with thinking about how to improve intervention development and implementation planning. However, they should not be used to cast doubt unfairly on null or negative effect findings, particularly if contextual issues were not considered before the findings were known. |
| 7. | This argument can be used to imply that the intervention | Forecasting delayed effects may be reasonable if there are good theoretical or empirical grounds to justify it (e.g., observed effects on proposed mediators). When these are not present, it can cast doubt on null results unfairly, particularly in the absence of the means or intention of investigating longer-term effects. |
Influences on what happens to an intervention following a null or negative effect trial
| Factor | Continued or future delivery of the intervention in its current form is | Continued or future delivery of the intervention in its current form is |
|---|---|---|
| Stage of gestation | Intervention is mature and widely commissioned | Intervention is new or early in its development |
| Perceived importance | Well established and politically important | Lower profile and limited political importance |
| Implementation feasibility/acceptability | Easy to deliver well, liked by practitioners/users | Hard to deliver well, disliked by practitioners/users |
| Outcome(s) targeted | Considered important (is a potential threat to health) | Considered less important (not a threat to health) |
| Quality of trial design and conduct | Concerns about quality undermine confidence in results | Judged to be high quality and reliable |
| Pattern of outcome results | Somewhat inconsistent or inconclusive | Consistent and conclusive null/negative results |
| Context in which trial was conducted | Deemed significantly different to new/current context | Deemed to be similar to new/current context |
| Insight into reasons for the result | Explained by methodological or delivery issues | No reason to doubt or explain away the result |
| Nature of the control condition | Intervention of interest (I) vs. similar intervention (C) | Modification (I) vs. original intervention (C) |
| Evidence base for intervention | Multiple other trials with positive effects | No other trials, or other evidence equivocal |
| Wider evidence base | Similar interventions not obviously superior | Similar interventions show positive effects |
| Policy and practice imperatives | Need to do something, and nothing clearly superior | Some discretion about acting, or superior alternatives |
| Political and economic situation | Limited resources, and “better” alternatives cost more | Resources allow more effective but costly alternativea |
| Investment in the intervention | Strong psychological or financial investment | Weaker investment, permitting more detached stance |
| Outlook on evidence (particularly trials) | Skeptical about evidence-based practice and/or trials | Sympathetic towards evidence-based practice/trials |
aOnly applies to existing interventions, not those delivered solely in the context of a trial