| Literature DB >> 30151132 |
Abstract
Pseudoreplication is defined as the use of inferential statistics to test for treatment effects where treatments are not replicated and/or replicates are not statistically independent. It is a genuine but controversial issue in ecology particularly in the case of costly landscape-scale manipulations, behavioral studies where ethics or other concerns may limit sample sizes, ad hoc monitoring data, and the analysis of natural experiments where chance events occur at a single site. Here key publications on the topic are reviewed to illustrate the debate that exists about the conceptual validity of pseudoreplication. A survey of ecologists and case studies of experimental design and publication issues are used to explore the extent of the problem, ecologists' solutions, reviewers' attitudes, and the fate of submitted manuscripts. Scientists working across a range of ecological disciplines regularly come across the problem of pseudoreplication and build solutions into their designs and analyses. These include carefully defining hypotheses and the population of interest, acknowledging the limits of statistical inference and using statistical approaches including nesting and random effects. Many ecologists face considerable challenges getting their work published if accusations of pseudoreplication are made - even if the problem has been dealt with. Many reviewers reject papers for pseudoreplication, and this occurs more often if they haven't experienced the issue themselves. The concept of pseudoreplication is being applied too dogmatically and often leads to rejection during review. There is insufficient consideration of the associated philosophical issues and potential statistical solutions. By stopping the publication of ecological studies, reviewers are slowing the pace of ecological research and limiting the scope of management case studies, natural events studies, and valuable data available to form evidence-based solutions. Recommendations for fair and consistent treatment of pseudoreplication during writing and review are given for authors, reviewers, and editors.Entities:
Keywords: Bayesian statistics; P‐values; confounded effects; hypothesis formation; nesting; peer review; random effects; scientific publication; statistical population
Year: 2015 PMID: 30151132 PMCID: PMC6102510 DOI: 10.1002/ece3.1782
Source DB: PubMed Journal: Ecol Evol ISSN: 2045-7758 Impact factor: 2.912
Previous authors’ solutions to the problem of pseudoreplication and potential issues and pitfalls. Many of these suggestions will match our own, and the issue is therefore much more a function of some researchers’ and editors’ attitudes and perceptions of the issue
| Proposed solution | References | Issues |
|---|---|---|
| Authors should clearly articulate potential confounding effects. Be explicit about experimental designs | Schank and Koehnle ( | Gives ammunition to reviewers who often seem to dislike studies without “perfect” designs |
| Compare a single treatment with multiple controls | Oksanen ( | Many statistical tests require variance estimates for treatment and control |
| Where site and treatment are confounded, examine magnitudes of difference between treatment and control areas before and after the experiment | Oksanen ( | Requires information on predisturbance conditions |
| Utilize meta‐analysis to investigate cross‐study comparisons | Hargrove and Pickering ( | File drawer problem and bias against pseudoreplication means many observational studies and negative results are not published |
| Accompany presentation of all results with inferential statistics | Oksanen ( | Care needed to avoid over interpretation if, for example, sites and treatments are confounded |
| Use inferential statistics to assess the “reliability” of descriptive statistics | Cottenie and De Meester ( | Care needed to avoid over interpretation |
| Focus on effect sizes, “how different the two statistical populations must be,” and divergence/convergence of temporal trends | Oksanen ( | Editors and reviewers (still) routinely demand |
| Avoid pooling of observations and instead use multilevel modeling as a statistical solution. | Waller et al. ( | More complex statistical methods needed which require expertise |
| Incorporate turnover‐by‐distance relationships and environmental data into their analyses to assess potential for spurious detection of significant differences. | Ramage et al. ( | Complex statistical analyses needed |
| Utilize Bayesian statistics | Oksanen ( | You'll need to understand Bayesian statistics first! |
| Carefully consider and clearly state the statistical inferences that can be drawn from data sets | Ramage et al. ( | Provides ammunition to those reviewers and editors looking for a reason to reject papers. Traditional journals like to maintain high rejection rates. |
| Pseudoreplication often related to confounded effects that require careful interpretation | Schank and Koehnle ( | See above |
| Explicitly state the limited scope of the results | Cottenie and De Meester ( | No one seems to want to publish “case studies” |
| Permit use of “normic statements” that hypothesize about what would normally occur given the results from a particular case study or statistical test | Hargrove and Pickering ( | Reviewers seem to dislike speculation. It would be better to phrase normic statements as new hypotheses to test |
| Substitute statistical inference for ecological inference | Hargrove and Pickering ( | Requires acknowledgment of precisely what statistical tests are comparing (e.g., site vs. treatment differences) |
| Allow publication of studies without inferential statistics | Hurlbert ( | Prevents authors from examining the extent to which observed differences are meaningful. Editors and reviewers (still) routinely demand |
| Avoid use of term pseudoreplication during review and instead specifically describe perceived statistical problems | Oksanen ( | No argument from us here! |
| Do not automatically reject “experiments” where there is no treatment replication | Hurlbert ( | We couldn't agree more! |
| Pseudoreplication should be taken into account when applicable | Cottenie and De Meester ( | Allows continued use of an imprecise term and doesn't encourage reviewers to specify exact statistical issues |
Categories of pseudoreplication problem identified in the questionnaire and the frequency with which respondents described them
| Landscape‐scale treatments/monitoring (including manipulations of forest stand structure) | 10 |
| Nested designs with insufficient replication at site level | 9 |
| Wildlife behavior/physiology (including repeated measures on a small number of individuals) | 9 |
| Confounded site/stand and treatment (including multisite vegetation chronosequences) | 8 |
| Demography and disease and – what is the appropriate analysis level site, plot, or individual? | 8 |
| Exclosures at a single site (including grazing and irrigation studies) | 6 |
| Aquatic ecology + hydrology ‐ unreplicated ponds/lakes/watersheds | 5 |
| Fire behavior and effects (including studies of individual wildfires) | 5 |
| Single‐site case studies or phenomena limited to one location | 5 |
| Spatial autocorrelation | 3 |
| Repeated measures of vegetation change (including studies on a single relevé) | 2 |
Figure 1Hypothetical example of two experimental designs examining the effects of some form of disturbance (e.g., wildfire, grazing) on vegetation structure at two different elevations. Design A is a formal experiment, whereas Design B is a researcher's response to a natural (i.e., unplanned by the researcher) event. Dotted lines mark the perimeter of the disturbances which could be, for example, a series of experimental fires in A and a wildfire in B or a number of grazing exclosures in A vs. a landscape‐scale fence in B. Assuming that soils, slope, and aspect are more or less homogenous, at least within each studied elevational band, are the results of Design A more ecologically meaningful than those in B? Which design more adequately captures the ecological reality of wildfire or landscape‐scale alterations to grazing management? We would argue that Design A might actually sacrifice ecological reality for statistical independence as, for example, small fires cannot mimic a wildfire event and small grazing exclosures do not allow natural movement of grazers across landscapes.
| Resource issues | Statistical solutions |
|---|---|
| “The issue was always resulting from the balance between what is | “They [the reviewers] were too focused on the possibility and |
| reasonably possible and what is ideal… People scream | effects of pseudoreplication than our approach to dealing with it” |
| pseudoreplication when it's not pseudoreplication” | |
| “Hurlburt did us all a disservice when he said that statistics can't | |
| “…pseudoreplication is an issue that's been blown way out of | be used when pseudoreplication is present. They can but what |
| proportion. The real issue is how you interpret your results and | they tell you is something that warrants careful interpretation.” |
| then report them. How do you replicate things like marshes, forest | |
| patches? You have to say, what I found is true for this forest and | “Pseudoreplication is a bogus term for a poorly nested design. The |
| then the question is how representative that forest is of all the forests in the area” | Hurlbert publication is one of the most pernicious publications in all of ecology.” |
| “Ultimately all field work is pseudoreplicated, depending upon | “Pseudoreplication is just a question of correct model |
| scale. I have been criticized because all my work occurred in 1 | specification. If the model correctly reflects the sampling design, |
| estuary, 1 only in the Gulf of Mexico” | then the issue becomes one of parameter estimation and |
| potential parameter nonidentifiability.” | |
| “Often there is simply not the funding to conduct landscape scale | |
| experiments without pseudoreplication” | “I would say “generalized linear multilevel models” but, yes, |
| basically random effects” | |
| “Replication is not always possible in ecology. This is particularly | |
| true in restoration ecology when restored ecosystems are created | “I used multiple control sites, to at least differentiate the |
| at great expense and cannot always be replicated for the purpose | treatment area from multiple other sites.” |
| of scientific study. Sometimes we just have to study what is | |
| there!” | “Indicating that samples were taken at a distance greater than the |
| autocorrelation distance for many soil variables (from the | |
| “In many ways, it is very difficult to really meet the needs of | literature in the same ecosystem) and framing the results and |
| replication and even when we do it is often somewhat arbitrary. In | conclusions to this experiment design.” |
| many instances, the research we have done could be better | |
| referred to as ‘case studies,’ but then we'd have to pray a journal | |
| will accept that.” | |
| “Ecologists are so hung up on pseudoreplication, it's not even an | |
| issue for my hydrologist colleagues, in whose research | |
| pseudoreplication is often inevitable.” |