Literature DB >> 29091338

Analysis of family-wise error rates in statistical parametric mapping using random field theory.

Guillaume Flandin1, Karl J Friston1.   

Abstract

This technical report revisits the analysis of family-wise error rates in statistical parametric mapping-using random field theory-reported in (Eklund et al. []: arXiv 1511.01863). Contrary to the understandable spin that these sorts of analyses attract, a review of their results suggests that they endorse the use of parametric assumptions-and random field theory-in the analysis of functional neuroimaging data. We briefly rehearse the advantages parametric analyses offer over nonparametric alternatives and then unpack the implications of (Eklund et al. []: arXiv 1511.01863) for parametric procedures. Hum Brain Mapp, 40:2052-2054, 2019.
© 2017 The Authors Human Brain Mapping Published by Wiley Periodicals, Inc. © 2017 The Authors Human Brain Mapping Published by Wiley Periodicals, Inc.

Entities:  

Keywords:  family-wise error rate; random field theory; statistical parametric mapping; topological inference

Mesh:

Year:  2017        PMID: 29091338      PMCID: PMC6585687          DOI: 10.1002/hbm.23839

Source DB:  PubMed          Journal:  Hum Brain Mapp        ISSN: 1065-9471            Impact factor:   5.038


INTRODUCTION

Random field theory has been at the heart of statistical parametric mapping in neuroimaging—and its various implementations in academic software—for over two decades. With technical advances in data acquisition, its validity has been revisited every few years to ensure it is fit for purpose [Bennett et al., 2009; Hayasaka and Nichols, 2003; Hayasaka et al., 2004; Nichols, 2012; Pantazis et al., 2005; Woo et al., 2014; Worsley et al., 1996], particularly in relation to controlling family‐wise error. The statistical validity of procedures based on random field theory is important because random field theory offers an efficient and reproducible alternative to nonparametric testing. The advantages of parametric approaches over nonparametric approaches include the following: Parametric approaches are more efficient than their nonparametric counterpart by the Neyman–Pearson lemma. This follows because the most efficient test is based on the odds ratio inherent in parametric tests. This means that any nonparametric test can only be as efficient as a parametric test or less efficient. Parametric approaches are reproducible. In other words, one obtains the same result when repeating the analysis, unlike the P‐values based on samples of the null distribution used in nonparametric tests. Parametric approaches eschew the problem of complying with the exchangeability criteria of nonparametric procedures. These criteria make it difficult to apply nonparametric tests to data that have serial correlations or when using hierarchical models. Parametric approaches are computationally more efficient because they use distributional assumptions to eschew computationally intensive sampling from a null distribution. These advantages rest on distributional assumptions that, if violated, render parametric tests inexact. In these instances, one could consider using nonparametric tests. It is, therefore, important to ensure that parametric tests and random field theory are robust to any violations. The analyses reported by [Eklund et al., 2015] speak to this issue. So what conclusions can be drawn from these analyses?

A REVIEW OF THE EKLUND ET AL. SIMULATION RESULTS

Eklund et al. [2015] assess the family‐wise error rate using parametric and nonparametric tests and a variety of regressors to analyse (publicly available) resting state fMRI data from two sites. They manipulate a number of factors including: (i) inference based on peak height versus spatial extent; (ii) spatial extent inference based on high versus low cluster forming thresholds; (iii) under different levels of spatial smoothing for (iv) block versus event‐related regressors, using (v) one‐ and two‐sample t‐tests. In brief, they show that parametric inference based on peak height is well‐behaved and provides acceptable family‐wise error control. In contrast, parametric inference based on spatial extent is not valid when, and only when, a low cluster forming threshold is employed. This failure is well known and is consistent with random field theory: the null distribution for spatial extent is based on the Nosko conjecture that provides a distributional form for the spatial extent of a cluster [Friston et al., 1994]. The parameter of this distributional form is fixed using approximations to the expected number of maxima and the total volume above a threshold (see [Flandin and Friston, 2015] for a brief review). Crucially, both the distributional form for the spatial extent and the expected number of maxima (the Euler characteristic) are approximations that are only true in the limit of high thresholds (see fig. 1 in [Friston et al., 1994]). This means that tests based on spatial extent become inexact at low thresholds—as verified numerically by [Eklund et al., 2015].
Figure 1

Cluster‐level inference results for a two‐sample t‐test (two groups of 10 random subjects, repeated a thousand times) with the Beijing dataset using a cluster forming threshold of P = 0.001 (uncorrected) and the SPM12 software (r6685). Five levels of spatial smoothing were evaluated (4, 6, 8, 10 and 12 mm isotropic Gaussian kernels) with four different regressors (see [Eklund et al., 2015] for details).

Cluster‐level inference results for a two‐sample t‐test (two groups of 10 random subjects, repeated a thousand times) with the Beijing dataset using a cluster forming threshold of P = 0.001 (uncorrected) and the SPM12 software (r6685). Five levels of spatial smoothing were evaluated (4, 6, 8, 10 and 12 mm isotropic Gaussian kernels) with four different regressors (see [Eklund et al., 2015] for details). The effects of smoothing reported in [Eklund et al., 2015] are consistent with random field theory, which assumes a good lattice approximation to a continuous random field. This assumption means that the data have to be smoother than the size of voxels. In other words, increasing the smoothness will lead to more exact inference. Again, this is verified numerically by [Eklund et al., 2015]. The effect of one versus two‐sample t‐tests is slightly more difficult to interpret. This is because the authors used the same regressor for all subjects. Arguably, this was a mistake because any systematic fluctuation in resting state timeseries—that correlates with the regressor—will lead to significant one‐sample t‐tests against the null hypothesis of zero (e.g., magnetic equilibration effects). This effect is particularly marked for a regressor (called E1) that represents a fast and inefficiently estimated event‐related response every few seconds. Crucially, the nonparametric false positive rates are beyond the 95% confidence intervals. This means that this effect is actually expressed in the data over subjects and, therefore, fails as a model of the null behaviour. This failure is finessed when comparing parameter estimates between two groups using a two‐sample t‐test. In this instance, inferences based on spatial extent fall to acceptable family‐wise error rates. We confirmed this by reproducing the analysis (using the same data and regressors) reported in [Eklund et al., 2015] (see Fig. 1). These analyses use the close to original (3 mm) voxels sizes—as opposed to the upsampled (2 mm voxel) data as analysed in [Eklund et al., 2015].

CONCLUSION

In summary, we have taken the opportunity to comment on the relative utility of nonparametric and parametric procedures in classical inference: despite the simplicity and robustness of nonparametric tests, there are principled reasons for the predominance of parametric procedures. Having said this, nonparametric tests are extremely useful when validating parametric assumptions—and establishing robustness to their violations. This use of nonparametric tests was exemplified in a recent paper entitled “Cluster failure: Why fMRI inferences for spatial extent have inflated false‐positive rates” [Eklund et al., 2016]. The answer to the question posed by Eklund et al. is that their cluster tests failed because they violated the assumptions underlying analytic (random field theory) approximations to null distributions; in particular, the assumption that clusters are defined by reasonably high thresholds. A useful rule of thumb here is that if clusters have more than one peak, then the cluster forming threshold is probably too low. A sufficiently high threshold is usually guaranteed with the standard cluster forming threshold of P = 0.001 (uncorrected). More generally, the simulations reported in Eklund et al speak to the importance of revisiting the robustness of statistical tests as the nature of imaging data evolves and new researchers—who are not familiar with these foundational issues—enter the field.
  9 in total

1.  Validating cluster size inference: random field and permutation methods.

Authors:  Satoru Hayasaka; Thomas E Nichols
Journal:  Neuroimage       Date:  2003-12       Impact factor: 6.556

2.  Nonstationary cluster-size inference with random field and permutation methods.

Authors:  Satoru Hayasaka; K Luan Phan; Israel Liberzon; Keith J Worsley; Thomas E Nichols
Journal:  Neuroimage       Date:  2004-06       Impact factor: 6.556

3.  A unified statistical approach for determining significant signals in images of cerebral activation.

Authors:  K J Worsley; S Marrett; P Neelin; A C Vandal; K J Friston; A C Evans
Journal:  Hum Brain Mapp       Date:  1996       Impact factor: 5.038

4.  A comparison of random field theory and permutation methods for the statistical analysis of MEG data.

Authors:  Dimitrios Pantazis; Thomas E Nichols; Sylvain Baillet; Richard M Leahy
Journal:  Neuroimage       Date:  2005-04-01       Impact factor: 6.556

5.  The principled control of false positives in neuroimaging.

Authors:  Craig M Bennett; George L Wolford; Michael B Miller
Journal:  Soc Cogn Affect Neurosci       Date:  2009-12       Impact factor: 3.436

6.  Assessing the significance of focal activations using their spatial extent.

Authors:  K J Friston; K J Worsley; R S Frackowiak; J C Mazziotta; A C Evans
Journal:  Hum Brain Mapp       Date:  1994       Impact factor: 5.038

7.  Cluster failure: Why fMRI inferences for spatial extent have inflated false-positive rates.

Authors:  Anders Eklund; Thomas E Nichols; Hans Knutsson
Journal:  Proc Natl Acad Sci U S A       Date:  2016-06-28       Impact factor: 11.205

8.  Cluster-extent based thresholding in fMRI analyses: pitfalls and recommendations.

Authors:  Choong-Wan Woo; Anjali Krishnan; Tor D Wager
Journal:  Neuroimage       Date:  2014-01-08       Impact factor: 6.556

Review 9.  Multiple testing corrections, nonparametric methods, and random field theory.

Authors:  Thomas E Nichols
Journal:  Neuroimage       Date:  2012-04-12       Impact factor: 6.556

  9 in total
  67 in total

1.  Alterations of functional connectivity and intrinsic activity within the cingulate cortex of suicidal ideators.

Authors:  Henry W Chase; Anna Maria Segreti; Timothy A Keller; Vladimir L Cherkassky; Marcel A Just; Lisa A Pan; David A Brent
Journal:  J Affect Disord       Date:  2017-01-15       Impact factor: 4.839

2.  Differential responses to infant faces in relation to maternal substance use: An exploratory study.

Authors:  Helena J V Rutherford; Sarah W Yip; Patrick D Worhunsky; Sohye Kim; Lane Strathearn; Marc N Potenza; Linda C Mayes
Journal:  Drug Alcohol Depend       Date:  2019-12-16       Impact factor: 4.492

3.  FMRI Clustering in AFNI: False-Positive Rates Redux.

Authors:  Robert W Cox; Gang Chen; Daniel R Glen; Richard C Reynolds; Paul A Taylor
Journal:  Brain Connect       Date:  2017-04

4.  Dissecting the neurofunctional bases of intentional action.

Authors:  Laura Zapparoli; Silvia Seghezzi; Paola Scifo; Alberto Zerbi; Giuseppe Banfi; Marco Tettamanti; Eraldo Paulesu
Journal:  Proc Natl Acad Sci U S A       Date:  2018-06-27       Impact factor: 11.205

5.  Cluster failure or power failure? Evaluating sensitivity in cluster-level inference.

Authors:  Stephanie Noble; Dustin Scheinost; R Todd Constable
Journal:  Neuroimage       Date:  2019-12-15       Impact factor: 6.556

6.  Distinguishing neural correlates of context-dependent advantageous- and disadvantageous-inequity aversion.

Authors:  Xiaoxue Gao; Hongbo Yu; Ignacio Sáez; Philip R Blue; Lusha Zhu; Ming Hsu; Xiaolin Zhou
Journal:  Proc Natl Acad Sci U S A       Date:  2018-07-30       Impact factor: 11.205

7.  Brain networks underlying conscious tactile perception of textures as revealed using the velvet hand illusion.

Authors:  Nader Rajaei; Naoya Aoki; Haruka K Takahashi; Tetsu Miyaoka; Takanori Kochiyama; Masahiro Ohka; Norihiro Sadato; Ryo Kitada
Journal:  Hum Brain Mapp       Date:  2018-08-10       Impact factor: 5.038

8.  Neural systems of cognitive demand avoidance.

Authors:  Ceyda Sayalı; David Badre
Journal:  Neuropsychologia       Date:  2018-06-23       Impact factor: 3.139

9.  Cluster-level statistical inference in fMRI datasets: The unexpected behavior of random fields in high dimensions.

Authors:  Ravi Bansal; Bradley S Peterson
Journal:  Magn Reson Imaging       Date:  2018-02-03       Impact factor: 2.546

10.  Neural correlates of taste reward value across eating disorders.

Authors:  Aviva K Olsavsky; Megan E Shott; Marisa C DeGuzman; Guido K W Frank
Journal:  Psychiatry Res Neuroimaging       Date:  2018-08-18       Impact factor: 2.376

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.