| Literature DB >> 35873757 |
Damien R Farine1,2,3, Gerald G Carter4,5.
Abstract
Permutation tests are widely used to test null hypotheses with animal social network data, but suffer from high rates of type I and II error when the permutations do not properly simulate the intended null hypothesis.Two common types of permutations each have limitations. Pre-network (or datastream) permutations can be used to control 'nuisance effects' like spatial, temporal or sampling biases, but only when the null hypothesis assumes random social structure. Node (or node-label) permutation tests can test null hypotheses that include nonrandom social structure, but only when nuisance effects do not shape the observed network.We demonstrate one possible solution addressing these limitations: using pre-network permutations to adjust the values for each node or edge before conducting a node permutation test. We conduct a range of simulations to estimate error rates caused by confounding effects of social or non-social structure in the raw data.Regressions on simulated datasets suggest that this 'double permutation' approach is less likely to produce elevated error rates relative to using only node permutations, pre-network permutations or node permutations with simple covariates, which all exhibit elevated type I errors under at least one set of simulated conditions. For example, in scenarios where type I error rates from pre-network permutation tests exceed 30%, the error rates from double permutation remain at 5%.The double permutation procedure provides one potential solution to issues arising from elevated type I and type II error rates when testing null hypotheses with social network data. We also discuss alternative approaches that can provide robust inference, including fitting mixed effects models, restricted node permutations, testing multiple null hypotheses and splitting large datasets to generate replicated networks. Finally, we highlight ways that uncertainty can be explicitly considered and carried through the analysis.Entities:
Keywords: animal social networks; hypothesis testing; permutation tests; social behaviour; social network analysis
Year: 2021 PMID: 35873757 PMCID: PMC9297917 DOI: 10.1111/2041-210X.13741
Source DB: PubMed Journal: Methods Ecol Evol Impact factor: 8.335
FIGURE 1Overview of the double permutation method. First, pre‐network permutations (1) generate a null distribution of expected metric values for a unit of interest (e.g. a node's degree or an edge's weight) alongside each unit's observed metric (2). Next, to remove nuisance effects for each unit, the median of expected metric values is subtracted from the observed metric value, yielding a corrected metric value called the ‘deviation score’ (3). Then, to generate a corrected test statistic, we fit a model with the deviation scores (4). To generate a p‐value, a nonparametric node permutation test can then be used to compare the corrected and expected test statistics (5). Alternatively, an appropriate parametric model could be used to replace (4) and (5)
Propensity for different permutation tests to yield errors or detect real effects when using regression models to test hypotheses on networks collected using gambit‐of‐the‐group data (model 1). Table shows the proportion of statistically significant results for an effect of a trait on degree under three sets of scenarios. When no effects are present, the expected proportion of significant results should be 5%. When a social effect is present, most results should be significant. When a spatial confound is present, the proportion of significant results should again approach 5%. Controlled node permutations include the number of observations in the models plus individuals' most common location in the spatial confound condition. Figures S2–S4 show how the proportion of significant results is affected by the number of observations and the number of nodes in the network
| No effects (should be low) | Social effect only (should be high) | Spatial confound (should be low) | |
|---|---|---|---|
| Degree | |||
| Node permutation ( | 4.8% | 88.4% | 88.4% |
| Controlled node permutation ( | 4.8% | 90.4% | 47.0% |
| Pre‐network permutation ( | 26.0% | 90.2% | 23.9% |
| Pre‐network permutation ( | 13.7% | 65.3% | 29.3% |
| Controlled pre‐network permutation | 26.0% | 89.7% | 23.9% |
| Double permutation | 4.7% | 89.7% | 10.8% |
| Eigenvector centrality | |||
| Node permutation ( | 5.1% | 93.2% | 92.5% |
| Controlled node permutation ( | 5.1% | 93.6% | 63.6% |
| Pre‐network permutation ( | 34.1% | 95.3% | 39.7% |
| Pre‐network permutation ( | 15.2% | 76.6% | 22.3% |
| Controlled pre‐network permutation | 34.3% | 95.0% | 23.1% |
| Double permutation | 4.9% | 93.4% | 33.6% |
| Betweenness | |||
| Node permutation ( | 4.6% | 86.3% | 86.5% |
| Controlled node permutation ( | 4.9% | 85.6% | 72.9% |
| Pre‐network permutation ( | 20.2% | 74.5% | 69.3% |
| Pre‐network permutation ( | 11.2% | 72.5% | 69.7% |
| Controlled pre‐network permutation | 19.9% | 74.3% | 65.6% |
| Double permutation | 4.9% | 70.9% | 67.0% |
Propensity for permutation tests to produce type I and type II errors from datasets simulating focal sampling (model 2). Simulations comprise four scenarios: (a) females and males have identical social phenotypes and are observed equally, (b) females are more social and both sexes are observed equally, (c) females and males have identical social phenotypes but observations are biased towards males (20% of observations of females are missed), and (d) females are more social but observations are biased towards males (20% of observations of females are missed)
| No observation bias | Observation bias (‘nuisance’ effect) | |||
|---|---|---|---|---|
|
Phenotypes equal (Type I errors) |
Females more social (Type II errors) |
Phenotypes equal (Type I errors) |
Females more social (Type II errors) | |
| Degree | ||||
| Node permutation ( | 5.0% | 1.2% | 60.4% | 37.8% |
| Controlled node permutation ( | 5.6% | 20.0% | 68.2% | 10.4% |
| Pre‐network permutation ( | 37.8% | 5.2% | 34.2% | 7.4% |
| Pre‐network permutation ( | 30.8% | 69.0% | 60.0% | 36.2% |
| Controlled pre‐network permutation | 39.0% | 24.2% | 69.6% | 12.4% |
| Double permutation | 5.2% | 6.6% | 7.0% | 6.0% |
| Eigenvector centrality | ||||
| Node permutation ( | 4.8% | 2.4% | 55.0% | 31.6% |
| Controlled node permutation ( | 7.0% | 2.4% | 71.4% | 7.0% |
| Pre‐network permutation ( | 45.2% | 1.2% | 43.0% | 3.0% |
| Pre‐network permutation ( | 23.4% | 74.4% | 64.4% | 34.6% |
| Controlled pre‐network permutation | 38.8% | 14.0% | 79.2% | 5.2% |
| Double permutation | 4.8% | 4.2% | 6.2% | 7.0% |
| Betweenness | ||||
| Node permutation ( | 5.8% | 8.6% | 67.0% | 52.8% |
| Controlled node permutation ( | 4.6% | 34.4% | 33.6% | 93.2% |
| Pre‐network permutation ( | 17.0% | 69.2% | 19.0% | 84.6% |
| Pre‐network permutation ( | 15.9% | 68.7% | 18.3% | 79.9% |
| Controlled pre‐network permutation | 17.2% | 69.4% | 19.8% | 82.0% |
| Double permutation | 4.6% | 86.2% | 5.2% | 26.2% |
Propensity for permutation tests to produce type I and type II errors regarding kinship effects from simulated datasets with confounding social effects, that is, nonrandom social structure (model 3). Table shows the type I error rates in simulations where the social effect is a confound (i.e. strong associations are not linked to kinship), and estimated type II error rates in simulations where the social effect corresponds to the hypothesis being tested (i.e. strong associations are linked to kinship). Figures S5 and S6 show how the proportion of significant results is affected by the number of observations and the number of nodes in the network. While pre‐network permutations appear to outperform other approaches with respect to Type II errors, this is likely because they are also more sensitive to weak effects in small networks, which are likely to correspond to type I errors rather than correctly identifying a true effect (see Figure S6)
|
Kinship ≠ Associations (Type I errors) |
Kinship ∝ Associations (Type II errors) | |
|---|---|---|
| Node permutation ( | 5.1% | 12.5% |
| Controlled node permutation ( | 4.9% | 12.0% |
| Pre‐network permutation ( | 18.6% | 6.2% |
| Pre‐network permutation ( | 3.0% | 79.1% |
| Double permutation | 5.2% | 16.5% |
Recommendations from simulations for choice of permutation tests. In the absence of nuisance effects, or when weak nuisance effects can effectively be controlled in a node permutation (e.g. using restricted node permutations), then either node permutations or double permutations are likely to provide robust inference for most local network metrics (e.g. degree, eigenvector centrality) or for relational data (tests on edge weights). In the presence of nuisance effects, double permutations tests are generally recommended, except for betweenness
| Metric | Relational (edge weight) | |||
|---|---|---|---|---|
| Degree centrality | Eigenvector centrality | Betweenness centrality | ||
| No clear nuisance effects | Node or double | Node or double | Node | Node or double |
| Nuisance effects expected | Double | Double | (Unclear) | Double |