| Literature DB >> 30964856 |
Sophie K Piper1,2, Ulrike Grittner1,2, Andre Rex3,4, Nico Riedel5, Felix Fischer6, Robert Nadon5,7,8, Bob Siegerink3, Ulrich Dirnagl3,5,9,10,11.
Abstract
The need for replication of initial results has been rediscovered only recently in many fields of research. In preclinical biomedical research, it is common practice to conduct exact replications with the same sample sizes as those used in the initial experiments. Such replication attempts, however, have lower probability of replication than is generally appreciated. Indeed, in the common scenario of an effect just reaching statistical significance, the statistical power of the replication experiment assuming the same effect size is approximately 50%-in essence, a coin toss. Accordingly, we use the provocative analogy of "replicating" a neuroprotective drug animal study with a coin flip to highlight the need for larger sample sizes in replication experiments. Additionally, we provide detailed background for the probability of obtaining a significant p value in a replication experiment and discuss the variability of p values as well as pitfalls of simple binary significance testing in both initial preclinical experiments and replication studies with small sample sizes. We conclude that power analysis for determining the sample size for a replication study is obligatory within the currently dominant hypothesis testing framework. Moreover, publications should include effect size point estimates and corresponding measures of precision, e.g., confidence intervals, to allow readers to assess the magnitude and direction of reported effects and to potentially combine the results of initial and replication study later through Bayesian or meta-analytic approaches.Entities:
Mesh:
Year: 2019 PMID: 30964856 PMCID: PMC6456162 DOI: 10.1371/journal.pbio.3000188
Source DB: PubMed Journal: PLoS Biol ISSN: 1544-9173 Impact factor: 8.029
Fig 1Power of replication experiment depending on the expected effect size and sample size.
Colored numerical values refer to our original experiment. Data of this figure can be found in S1 Data, and the figure can be explored further under s-quest.bihealth.org/power_replication/.
Fig 2Results of the “replication” experiment.
Screenshots of the coin flip experiment: (A) blind selection of coin and (B) flipping the coin (C) resulting in heads.
Attributes and applications of different methods of replication.
| Method of replication/ | Coin flip replication | Exact replication (same design, same sample size) | Exact replication with increased sample size (e.g., 2.5× sample size of initial study) | Conceptual replication (meaningful alterations to design, varying sample size) |
|---|---|---|---|---|
| Can identify technical mistakes in initial experiment | no | yes | yes | maybe |
| Can be used to reduce false inference on treatment effects | no | maybe | yes | maybe |
| Can provide information on robustness | no | no | no | yes |
| Can be used for meta-analyses | no | yes | yes | maybe |