Literature DB >> 30670959

An Empirical Comparison of Meta- and Mega-Analysis With Data From the ENIGMA Obsessive-Compulsive Disorder Working Group.

Premika S W Boedhoe1,2, Martijn W Heymans3, Lianne Schmaal4,5, Yoshinari Abe6, Pino Alonso7,8,9, Stephanie H Ameis10,11, Alan Anticevic12, Paul D Arnold13,14, Marcelo C Batistuzzo15, Francesco Benedetti16, Jan C Beucke17, Irene Bollettini16, Anushree Bose18, Silvia Brem19, Anna Calvo20, Rosa Calvo8,21, Yuqi Cheng22, Kang Ik K Cho23, Valentina Ciullo24,25, Sara Dallaspezia16, Damiaan Denys26,27, Jamie D Feusner28, Kate D Fitzgerald29, Jean-Paul Fouche30, Egill A Fridgeirsson26, Patricia Gruner12, Gregory L Hanna29, Derrek P Hibar31, Marcelo Q Hoexter15, Hao Hu32, Chaim Huyser33,34, Neda Jahanshad35, Anthony James36, Norbert Kathmann17, Christian Kaufmann17, Kathrin Koch37,38, Jun Soo Kwon39,40, Luisa Lazaro8,21,41,42, Christine Lochner43, Rachel Marsh44,45, Ignacio Martínez-Zalacaín7, David Mataix-Cols46, José M Menchón7,8,9, Luciano Minuzzi47, Astrid Morer8,21,41, Takashi Nakamae6, Tomohiro Nakao48, Janardhanan C Narayanaswamy18, Seiji Nishida6, Erika L Nurmi28, Joseph O'Neill28, John Piacentini28, Fabrizio Piras24, Federica Piras24, Y C Janardhan Reddy18, Tim J Reess37,38, Yuki Sakai6,49, Joao R Sato50, H Blair Simpson44,51, Noam Soreni52, Carles Soriano-Mas7,8,53, Gianfranco Spalletta24,54, Michael C Stevens55,56, Philip R Szeszko57,58, David F Tolin55,59, Guido A van Wingen26, Ganesan Venkatasubramanian18, Susanne Walitza19, Zhen Wang32,60, Je-Yeon Yun35,39, Paul M Thompson31, Dan J Stein30, Odile A van den Heuvel1,2, Jos W R Twisk3.   

Abstract

Objective: Brain imaging communities focusing on different diseases have increasingly started to collaborate and to pool data to perform well-powered meta- and mega-analyses. Some methodologists claim that a one-stage individual-participant data (IPD) mega-analysis can be superior to a two-stage aggregated data meta-analysis, since more detailed computations can be performed in a mega-analysis. Before definitive conclusions regarding the performance of either method can be drawn, it is necessary to critically evaluate the methodology of, and results obtained by, meta- and mega-analyses.
Methods: Here, we compare the inverse variance weighted random-effect meta-analysis model with a multiple linear regression mega-analysis model, as well as with a linear mixed-effects random-intercept mega-analysis model, using data from 38 cohorts including 3,665 participants of the ENIGMA-OCD consortium. We assessed the effect sizes and standard errors, and the fit of the models, to evaluate the performance of the different methods.
Results: The mega-analytical models showed lower standard errors and narrower confidence intervals than the meta-analysis. Similar standard errors and confidence intervals were found for the linear regression and linear mixed-effects random-intercept models. Moreover, the linear mixed-effects random-intercept models showed better fit indices compared to linear regression mega-analytical models. Conclusions: Our findings indicate that results obtained by meta- and mega-analysis differ, in favor of the latter. In multi-center studies with a moderate amount of variation between cohorts, a linear mixed-effects random-intercept mega-analytical framework appears to be the better approach to investigate structural neuroimaging data.

Entities:  

Keywords:  IPD meta-analysis; MRI; linear mixed-effect models; mega-analysis; neuroimaging

Year:  2019        PMID: 30670959      PMCID: PMC6331928          DOI: 10.3389/fninf.2018.00102

Source DB:  PubMed          Journal:  Front Neuroinform        ISSN: 1662-5196            Impact factor:   4.081


Introduction

Data pooling across individual studies has the potential to significantly accelerate progress in brain imaging (Van Horn et al., 2001), as demonstrated by large-scale neuroimaging initiatives, such as the ENIGMA (Enhanced NeuroImaging Genetics through Meta-Analysis) consortium (Thompson et al., 2014). The most immediate advantage of data pooling is increased power due to the larger number of subjects available for analysis. Data pooling across multiple centers worldwide can also lead to a more heterogeneous and potentially representative participant sample. Large-scale studies are well-powered to distinguish consistent, generalizable findings from false positives that emerge from smaller-sampled studies. The participation of many experts may also lead to a more balanced interpretation, wider endorsement of the conclusions by others, and greater dissemination of results (Stewart, 1995). An aggregate data meta-analysis is the most conventional approach, where summary results, such as effect size estimates, standard errors, and confidence intervals, are extracted from primary published studies and then synthesized to estimate the overall effect for all the studies combined (de Bakker et al., 2008). This approach is relatively quick and inexpensive, but often prone to selective reporting in primary studies, publication bias, low power to detect interaction effects and lack of harmonization of data processing and analysis methods among the included studies. To overcome these issues, collaborative groups are increasingly collating individual-participant data (IPD) from multiple studies to jointly analyze the individual-level data in a meta-analysis of IPD (Stewart, 1995). The IPD approach allows standardization of processing protocols and statistical analyses, culminating in study results not provided by the individual publications. This approach also allows modeling of interaction effects within the studies. Given these advantages, the IPD approach is currently the gold standard. There are two competing statistical approaches for IPD meta-analysis: a two-stage or a one-stage approach (Thomas et al., 2014). In the two-stage approach, the first step includes analyzing the IPD from each study separately, to obtain aggregate (summary) data (e.g., effect size estimates and confidence intervals). The second step includes using standard meta-analytical techniques, such as a random effects meta-analysis model. The alternative one-stage approach analyzes all IPD in one statistical model while accounting for clustering among patients in the same study, to estimate an overall effect. Throughout this manuscript, the one-stage IPD approach is referred to as mega-analysis, while the two-stage approach is referred to as meta-analysis. Some methodologists claim that a mega-analysis can be superior to meta-analysis. The comprehensive evaluation of missing data and greater flexibility in the control of confounders at the level of individual patients and specific studies are significant advantages of a mega-analytical approach. Mega-analyses have also been recommended as they avoid the assumptions of within-study normality and known within-study variances, which are especially problematic with smaller samples (Debray et al., 2013). Despite these advantages, mega-analysis requires homogeneous data sets and the establishment of a common centralized database. The latter criterion is time-consuming since cleaning, checking, and re-formatting the various data sets adds to the time and costs of performing mega-analyses. Obtaining IPD may also be challenging and limited by the terms of the informed consent or other data sharing constraints within each study. These are the main reasons why researchers often prefer meta-analysis using summary statistics. Additionally, meta-analysis allows for analyses of individual studies to account for local population substructure and study-specific covariates that may be better dealt with within each study. While each method has its own advantages and limitations, researchers still debate which method is superior for tackling different types of questions [see (Stewart and Tierney, 2002; Burke et al., 2017) for reviews on advantages and disadvantages of each approach]. Brain imaging communities focusing on different diseases have started collaborating to perform well-powered meta- and mega-analyses. In the largest studies to date on the neural correlates of OCD, the authors of the ENIGMA-OCD consortium (Boedhoe et al., 2017a, 2018) conducted a mega-analysis, pooling individual participant-level data from more than 25 research institutes worldwide, as well as a meta-analysis by combining summary statistic results from the independent sites. The meta- and mega-analyses revealed comparable findings of subcortical abnormalities in OCD (Boedhoe et al., 2017a), but the mega-analytical approach seemed more sensitive for detecting subtle cortical abnormalities (Boedhoe et al., 2018). Before definitive conclusions regarding the performance of either method can be drawn, it is necessary to critically evaluate the results obtained by various approaches for meta- and mega-analyses. Herein, we use data from the ENIGMA-OCD consortium to compare results obtained by meta- and mega-analyses. Specifically, we applied the inverse variance weighted random-effect meta-analysis model and the multiple linear regression mega-analysis model as used in the aforementioned studies (Boedhoe et al., 2017a, 2018). In addition, we compared findings from these models to those detected with a linear mixed-effects random-intercept mega-analytical model. Effect sizes and standard error estimates, and (where possible) model fit were used to evaluate which of the methods performs best.

Methods

Samples

The ENIGMA-OCD working group includes 38 data sets from 27 international research institutes with neuroimaging and clinical data from OCD patients and typically developing healthy control subjects, including both children and adults (Boedhoe et al., 2018). We defined adults as individuals aged ≥18 years and children as individuals aged < 18 years. The split at the age of 18 followed from a natural selection of the age ranges used in these samples, as most samples used the age of 18 years as a cut-off for inclusion. Because our previous findings and the literature suggest differential effects between pediatric and adult samples, we performed separate analyses for adult and pediatric data [for demographics and further details on the samples, see (Boedhoe et al., 2018)]. In total, we analyzed data from 3,665 participants including 1,905 OCD patients (407 children and 1,498 adults) and 1,760 control participants (324 children and 1,436 adults). All local institutional review boards permitted the use of measures extracted from the coded data for analyses.

Image Acquisition and Processing

Structural T1-weighted brain MRI scans were acquired and processed locally. For image acquisition parameters of each site, please see (Boedhoe et al., 2018). All cortical parcellations were performed with the fully automated segmentation software FreeSurfer, version 5.3 (Fischl, 2012), following standardized ENIGMA protocols to harmonize analyses and quality control procedures across multiple sites (see http://enigma.usc.edu/protocols/imaging-protocols/). Segmentations of 68 (34 left and 34 right) cortical gray matter regions based on the Desikan-Killiany atlas (Desikan et al., 2006) and two whole-hemisphere measures were visually inspected and statistically evaluated for outliers [see (Boedhoe et al., 2018) for further details on quality checking].

Statistical Framework

We examined differences between OCD patients and controls across samples by performing (1) an inverse variance weighted random-effects meta-analysis model; (2) a multiple linear regression mega-analysis model; and (3) a linear mixed-effects random-intercept mega-analysis model. Each of the 70 cortical regions of interest (68 regions and two whole-hemisphere averages) served as the outcome measure and a binary indicator of diagnosis as the predictor of interest. In the meta-analysis, all cortical thickness models were adjusted for age and sex (Im et al., 2008; Westlye et al., 2010), and all cortical surface area models were corrected for age, sex, and intracranial volume (Barnes et al., 2010; Ikram et al., 2012). In the mega-analysis all models were also adjusted for scanning center (cohort). The two mega-analytical frameworks are similar, but the models account differently for clustering of data within cohorts; linear regression with a dummy variable for each cohort and linear mixed-effects models (more efficiently) with only one variance parameter. Finally, all models were fit using the restricted maximum likelihood method [REML (Harville, 1977)]. The meta- and mega-analysis encompass intrinsically different statistics, including differences in approaches for dealing with missing data. E.g., the mega-analysis estimates one restricted maximum likelihood over the entire data set. This estimation contains information of each of the other cohorts. The first stage of the meta-analysis includes the estimation of a restricted maximum likelihood per cohort, making this method more vulnerable to missing outcome data. Therefore, we descriptively compared the meta- and mega-analyses by examining the confidence intervals and standard error estimates for the effect sizes assessed. In addition, the Bayesian information criterion (BIC) were used to evaluate which of the mega-analytical models performs better. A lower BIC indicates a better model fit. Throughout the manuscript, we report p < 0.001.

Meta-Analysis

We analyzed the IPD from each study to obtain aggregated summary data. Effect size estimates were calculated using Cohen's d, computed from the t-statistic of the diagnosis indicator variable from the regression models [(Nakagawa and Cuthill, 2007), equation 10]. All regression models and effect size estimates were fitted at each site separately. A final Cohen's d effect size estimate was obtained using an inverse variance-weighted random-effect meta-analysis model in R (metafor package, version 1.9-118). This meta-analytic framework enabled us to combine data from multiple sites and take the sample size of each cohort into account by weighing individual effect size estimates for the inverse variance per cohort.

Mega-Analysis

We pooled all IPD in one statistical model to perform mega-analyses and fitted the following models:

Linear Regression

The linear regression model included cohorts as dummy variables. Effect size estimates were calculated using the Cohen's d metric computed from the t-statistic of the diagnosis indicator variable from the regression models [(Nakagawa and Cuthill, 2007), equation 10].

Linear Mixed-Effects Model – Random-Intercept

Linear mixed-effects models are extensions of linear regression models and efficiently account for clustering of data within cohorts. By adding a random-intercept for cohort, the adjustment for the clustering of data within cohorts is performed with only one (variance) parameter, which reduces the number of estimated parameters (rather than estimating the intercept of each dummy variable separately as in the linear regression model described above). We used lme4 (linear mixed-effects analysis) package in R to perform the analyses. Effect size estimates were calculated using the Cohen's d metric computed from the t-values from the mixed-effects model [(Nakagawa and Cuthill, 2007), equation 22].

Results

The results of the meta-analysis and linear regression mega-analysis have been published previously (Boedhoe et al., 2018). In this paper, we added the linear mixed-effects random-intercept mega-analysis and statistically compared the various approaches. No significant differences (p < 0.001) in cortical thickness were observed in adult OCD patients (N = 1,498) compared to healthy controls (N = 1,436) (Supplementary Table S1). The meta-analysis did reveal a lower surface area of the transverse temporal cortex (Cohen's d −0.17) in OCD patients (Supplementary Table S2). No group differences in cortical thickness or surface area were observed in children with OCD (N = 407) compared to control children (N = 324) (Supplementary Tables S3, S4). Both the linear regression (Cohen's d −0.14) and the linear mixed-effects random-intercept (Cohen's d −0.11) models revealed significantly lower cortical thickness in bilateral inferior parietal cortices in adult OCD patients (N = 1,498) compared to healthy controls (N = 1,436) (Supplementary Table S5). Both models also showed significantly lower surface area (Cohen's d −0.16) in the left transverse temporal cortex in OCD patients (Supplementary Table S6). Both the linear regression (Cohen's d between −0.24 and −0.31) and the linear mixed-effects random-intercept (Cohen's d between −0.20 and −0.28) models revealed significantly thinner cortices in pediatric OCD patients (N = 407) compared with control children (N = 324) in the right superior parietal, left inferior parietal, and left lateral occipital cortices (Supplementary Tables S7). Neither model revealed significant group differences in cortical surface area (Supplementary Tables S8).

Comparing Meta- and Mega-Analysis

Effect Sizes

When looking at the magnitude and order of effect sizes we see the same pattern resulting from the meta-analysis and linear regression mega-analysis in both the pediatric (Supplementary Tables S3, S7) and adult (Supplementary Tables S1, S5) datasets, i.e., the magnitude and direction of effect of the effect sizes derived from the meta-analysis and linear regression mega-analysis were highly similar. The linear mixed-effects random-intercept mega-analysis also showed a similar pattern of results, but slightly smaller effect sizes (Table 1 and Supplementary Tables S5, S7).
Table 1

Effect size, confidence interval, and standard error estimates, and BIC values of the main findings.

Cohen's dStandard Error95% CIBIC
ADULT
Left inferior parietal cortexmeta-analysis−0.130.053−0.237 to −0.029
mega-analysis LR*−0.140.038−0.211 to −0.0639,693
cortical thicknessmega-analysis LMEri*−0.110.038−0.214 to −0.0659,638
Right inferior parietal cortexmeta-analysis−0.130.046−0.221 to −0.042
mega-analysis LR*−0.140.038−0.211 to −0.0649,799
mega-analysis LMEri*−0.100.038−0.213 to −0.0669,750
Left transverse temporal cortexcortical surface areameta-analysis*−0.170.038−0.243 to −0.095
mega-analysis LR*−0.160.037−0.238 to −0.09233,368
mega-analysis LMEri*−0.160.037−0.240 to −0.09533,263
PEDIATRIC
Right superior parietal cortexmeta-analysis−0.270.138−0.540 to −0.001
mega-analysis LR*−0.270.075−0.416 to −0.1212,645
mega-analysis LMEri*−0.210.075−0.415 to −0.1192,634
Left inferior parietal cortexcortical thicknessmeta-analysis−0.310.144−0.593 to −0.027
mega-analysis LR*−0.310.077−0.457 to −0.1542,634
mega-analysis LMEri*−0.280.077−0.455 to −0.1512,610
Left lateral occipital cortexmeta-analysis−0.260.095−0.445 to −0.071
mega-analysis LR*−0.260.075−0.404 to −0.1092,464
mega-analysis LMEri*−0.230.075−0.401 to −0.1062,444

LR, linear regression; LMEri, linear mixed-effects random-intercept model; CI, confidence interval; BIC = Bayesian information criterion.

Indicates significant group difference at a threshold of p < 0.001.

Effect size, confidence interval, and standard error estimates, and BIC values of the main findings. LR, linear regression; LMEri, linear mixed-effects random-intercept model; CI, confidence interval; BIC = Bayesian information criterion. Indicates significant group difference at a threshold of p < 0.001.

Standard Error and 95% Confidence Intervals

Overall, linear regression and linear mixed-effects random-intercept models showed lower standard errors and narrower confidence intervals than the meta-analysis. Similar standard errors and confidence intervals were found for the different mega-analysis models (Table 1 and Supplementary Tables S1–S8).

Goodness-of-Fit

The linear mixed-effects random-intercept models showed lower BIC values compared to the linear regression mega-analysis (Table 1 and Supplementary Tables S9–S12).

Discussion

The aim of this study was to evaluate different statistical methods for large-scale multi-center neuroimaging analyses. We empirically evaluated whether a meta-analysis provides results comparable to a mega-analysis and which analytical framework performs better. Clinical interpretation of the results can be found elsewhere (Boedhoe et al., 2017b, 2018). Although effect sizes were similar for the meta-analysis and linear regression mega-analysis, lower standard errors and narrower confidence intervals of both mega-analytical approaches compared to the meta-analysis suggest better performance of the mega-analytical approach over the meta-analytical approach. While the meta-analysis failed to detect cortical thickness differences in both the adult and pediatric samples, it did support the findings of the mega-analyses at a less stringent significance threshold (p < 0.05 uncorrected). As a second aim, we investigated which mega-analytical framework was superior. The BIC values indicated a better model fit of the linear mixed-effects random-intercept model compared to the linear regression mega-analytical model. Whereas, the linear regression model showed similar standard errors and confidence intervals to the linear mixed-effects random-intercept model, the latter fitted the data better. The effect sizes of the linear regression model appeared to be higher than those of the linear mixed-effects models, possibly indicating an overestimation of the effect of diagnosis. Indeed fixed-effects analyses (comparable to the linear regression models in our case) are reported to produce biased estimates or inflated type I error rates when pooled data includes cohorts with a small number of patients (Agresti and Hartzel, 2000; Kahan and Morris, 2012). Mathew and Nordstorm (2010) also suggested that a mega-analysis (one-stage approach) with a random intercept term might be slightly more precise than a meta-analysis (two-stage approach), which has a distinct intercept term per study (Mathew and Nordstorm, 2010). Taken together, our results suggest that the linear mixed-effects random-intercept mega-analysis model is the better approach for analyzing cortical gray matter data in a multi-center neuroimaging study. We also explored (data not shown) a linear mixed-effects random-intercept and random-slope mega-analytical approach, since the various cohorts might have shown differences in effects of diagnosis related to clinical heterogeneity between patient samples. However, for most of the regions of interest the model did not converge. These computational difficulties and convergence problems have been reported before (Debray et al., 2013). As a result, effect sizes, confidence intervals, standard errors, and BIC values could not be estimated accurately. Indeed previous literature has demonstrated that mega-analyses may produce downwardly biased coefficient estimates when an incorrect model is specified, for instance when random effects are wrongly assumed (Dutton, 2010). Note that including a random slope in the linear mixed-effects model might be valuable when large variance is present in the data between cohorts. Therefore, we recommend the following strategy: (1) run a mixed-effects model with a random-intercept to correct for clustering of participants within cohorts; (2) add a random-slope to correct for potential variance in effects between cohorts; (3) and perform a likelihood-ratio test to statistically compare both models. If the likelihood-ratio test is significant i.e., there is a better fit of the random-intercept random-slope model, this model is preferred over the random-intercept only model. If the likelihood-ratio test is not significant i.e., there is a better fit of the random-intercept only model, this model is preferred over the random-intercept random-slope model. Olkin and Sampson (1998) showed that for comparing treatments with respect to a continuous outcome in clinical trials, meta-analysis is equivalent to mega-analysis if the treatment effects and error variances are constant across trials. The equivalence has been extended even if the error variances are different across trials (Mathew and Nordstrom, 1999). Lin and Zeng theoretically and empirically showed asymptotic equivalence between meta- and mega-analyses when the effect sizes are the same for all studies (Lin and Zeng, 2010a,b). The different cohorts in our study did not all show similar effect sizes and error variances, possibly explaining why we did not find the meta- and mega-analyses to be equivalent. In practice, effect sizes and error variances vary across studies more often than not. Moreover, these authors (Lin and Zeng, 2010a) focused on a fixed-effects meta-analysis rather than a random-effects meta-analysis which is carried out in the current study. A fixed-effect model only takes into account the random error within cohorts, whereas the random-effect model also takes into account the random error between cohorts (Borenstein et al., 2010). Not taking into account the random error between different cohorts in neuroimaging data, for example, may lead to potentially misleading conclusions. More comprehensive simulation studies may be performed to assess theoretical differences in the results of meta- and mega-analyses. Such simulation studies covering various scenarios regarding varying effect sizes and error variances would strengthen our findings. Conclusions of meta-analyses are often used to guide health care policy and to make decisions regarding the management of individual patients. Thus, it is important that the conclusions of meta-analyses are valid. Although the two approaches (meta- and mega-analysis) often produce similar results, sometimes clinical and/or statistical conclusions are affected (Burke et al., 2017). We agree with Burke et al. (2017) and Debray et al. (2013) that when planning IPD analyses in a multi-center setting, the choice and implementation of a mega-analysis (one-stage approach) or meta-analysis (two-stage approach) method should be pre-specified, as occasionally they lead to different conclusions. Standardized statistical guidelines addressing the best approach, such as those mentioned in Burke et al. (2017), would be beneficial in this area. For example, meta-analysis (two-stage approach) or mega-analysis (one-stage approach) may be more suitable, depending on outcome types (continuous, binary of time-to-event). In a multi-center study including multiple small sample cohorts, a mega-analysis (one-stage approach) is preferred, as it avoids the use of approximate normal sampling distributions, known within-study variances, and continuity corrections that plague mega-analysis (two-stage approach) with an inverse variance weighting. Additionally, any mega-analysis (one-stage approach) should account for the clustering of participants within cohorts, ideally by including a random-intercept term for cohort. If the effect sizes of the separate studies are expected to vary greatly, it should be investigated whether adding a random-slope to the model is beneficial. For further details about choosing an appropriate method for a multi-center study we recommend Burke et al. (2017). To our knowledge, this is the first report investigating the utility of meta- vs. mega-analyses for multi-center structural neuroimaging data. The validity of our findings is limited to cortical gray matter measures. Therefore, they may not be generalized to all other brain measures. Nevertheless, our findings show that in the case of cross-sectional structural neuroimaging data a mega-analysis performs better than a meta-analysis. In a multi-center study with a moderate amount of variation between cohorts, a linear mixed-effects random-intercept mega-analytical framework seems to be the better approach to investigate structural neuroimaging data. We urge researchers worldwide to join forces by sharing data with the goal of elucidating biomedical problems that no group could address alone.

Ethics Statement

All subjects gave written informed consent in accordance with the Declaration of Helsinki. All local institutional review boards permitted the use of measures extracted from the coded data for meta- and mega-analysis.

Author Contributions

PB, MH, LS, OvdH, and JT contributed to the conception and design of the study. OvdH and JT contributed equally. PB organized the database. PB and MH performed the statistical analysis at the mega- and mega-analysis level. All other authors contributed to data processing and/or statistical analysis at site level. PB wrote the first draft of the manuscript. All other authors and members of the ENIGMA-OCD working group contributed to manuscript revision, read, and approved the manuscript.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
  26 in total

1.  Meta-analysis of age and actigraphy-assessed sleep characteristics across the lifespan.

Authors:  Marissa A Evans; Daniel J Buysse; Anna L Marsland; Aidan G C Wright; Jill Foust; Lucas W Carroll; Naina Kohli; Rishabh Mehra; Adam Jasper; Swathi Srinivasan; Martica H Hall
Journal:  Sleep       Date:  2021-09-13       Impact factor: 5.849

2.  Analysis of structural brain asymmetries in attention-deficit/hyperactivity disorder in 39 datasets.

Authors:  Merel C Postema; Martine Hoogman; Sara Ambrosino; Philip Asherson; Tobias Banaschewski; Cibele E Bandeira; Alexandr Baranov; Claiton H D Bau; Sarah Baumeister; Ramona Baur-Streubel; Mark A Bellgrove; Joseph Biederman; Janita Bralten; Daniel Brandeis; Silvia Brem; Jan K Buitelaar; Geraldo F Busatto; Francisco X Castellanos; Mara Cercignani; Tiffany M Chaim-Avancini; Kaylita C Chantiluke; Anastasia Christakou; David Coghill; Annette Conzelmann; Ana I Cubillo; Renata B Cupertino; Patrick de Zeeuw; Alysa E Doyle; Sarah Durston; Eric A Earl; Jeffery N Epstein; Thomas Ethofer; Damien A Fair; Andreas J Fallgatter; Stephen V Faraone; Thomas Frodl; Matt C Gabel; Tinatin Gogberashvili; Eugenio H Grevet; Jan Haavik; Neil A Harrison; Catharina A Hartman; Dirk J Heslenfeld; Pieter J Hoekstra; Sarah Hohmann; Marie F Høvik; Terry L Jernigan; Bernd Kardatzki; Georgii Karkashadze; Clare Kelly; Gregor Kohls; Kerstin Konrad; Jonna Kuntsi; Luisa Lazaro; Sara Lera-Miguel; Klaus-Peter Lesch; Mario R Louza; Astri J Lundervold; Charles B Malpas; Paulo Mattos; Hazel McCarthy; Leyla Namazova-Baranova; Rosa Nicolau; Joel T Nigg; Stephanie E Novotny; Eileen Oberwelland Weiss; Ruth L O'Gorman Tuura; Jaap Oosterlaan; Bob Oranje; Yannis Paloyelis; Paul Pauli; Felipe A Picon; Kerstin J Plessen; J Antoni Ramos-Quiroga; Andreas Reif; Liesbeth Reneman; Pedro G P Rosa; Katya Rubia; Anouk Schrantee; Lizanne J S Schweren; Jochen Seitz; Philip Shaw; Tim J Silk; Norbert Skokauskas; Juan C Soliva Vila; Michael C Stevens; Gustavo Sudre; Leanne Tamm; Fernanda Tovar-Moll; Theo G M van Erp; Alasdair Vance; Oscar Vilarroya; Yolanda Vives-Gilabert; Georg G von Polier; Susanne Walitza; Yuliya N Yoncheva; Marcus V Zanetti; Georg C Ziegler; David C Glahn; Neda Jahanshad; Sarah E Medland; Paul M Thompson; Simon E Fisher; Barbara Franke; Clyde Francks
Journal:  J Child Psychol Psychiatry       Date:  2021-03-22       Impact factor: 8.265

Review 3.  Wnt/β-catenin signaling in brain development and mental disorders: keeping TCF7L2 in mind.

Authors:  Joanna Bem; Nikola Brożko; Chaitali Chakraborty; Marcin A Lipiec; Kamil Koziński; Andrzej Nagalski; Łukasz M Szewczyk; Marta B Wiśniewska
Journal:  FEBS Lett       Date:  2019-06-30       Impact factor: 4.124

4.  Increased power by harmonizing structural MRI site differences with the ComBat batch adjustment method in ENIGMA.

Authors:  Joaquim Radua; Eduard Vieta; Russell Shinohara; Peter Kochunov; Yann Quidé; Melissa J Green; Cynthia S Weickert; Thomas Weickert; Jason Bruggemann; Tilo Kircher; Igor Nenadić; Murray J Cairns; Marc Seal; Ulrich Schall; Frans Henskens; Janice M Fullerton; Bryan Mowry; Christos Pantelis; Rhoshel Lenroot; Vanessa Cropley; Carmel Loughland; Rodney Scott; Daniel Wolf; Theodore D Satterthwaite; Yunlong Tan; Kang Sim; Fabrizio Piras; Gianfranco Spalletta; Nerisa Banaj; Edith Pomarol-Clotet; Aleix Solanes; Anton Albajes-Eizagirre; Erick J Canales-Rodríguez; Salvador Sarro; Annabella Di Giorgio; Alessandro Bertolino; Michael Stäblein; Viola Oertel; Christian Knöchel; Stefan Borgwardt; Stefan du Plessis; Je-Yeon Yun; Jun Soo Kwon; Udo Dannlowski; Tim Hahn; Dominik Grotegerd; Clara Alloza; Celso Arango; Joost Janssen; Covadonga Díaz-Caneja; Wenhao Jiang; Vince Calhoun; Stefan Ehrlich; Kun Yang; Nicola G Cascella; Yoichiro Takayanagi; Akira Sawa; Alexander Tomyshev; Irina Lebedeva; Vasily Kaleda; Matthias Kirschner; Cyril Hoschl; David Tomecek; Antonin Skoch; Therese van Amelsvoort; Geor Bakker; Anthony James; Adrian Preda; Andrea Weideman; Dan J Stein; Fleur Howells; Anne Uhlmann; Henk Temmingh; Carlos López-Jaramillo; Ana Díaz-Zuluaga; Lydia Fortea; Eloy Martinez-Heras; Elisabeth Solana; Sara Llufriu; Neda Jahanshad; Paul Thompson; Jessica Turner; Theo van Erp
Journal:  Neuroimage       Date:  2020-05-26       Impact factor: 6.556

5.  A multicountry perspective on gender differences in time use during COVID-19.

Authors:  Laura M Giurge; Ashley V Whillans; Ayse Yemiscigil
Journal:  Proc Natl Acad Sci U S A       Date:  2021-03-23       Impact factor: 11.205

6.  Toward Cumulative Cognitive Science: A Comparison of Meta-Analysis, Mega-Analysis, and Hybrid Approaches.

Authors:  Ezequiel Koile; Alejandrina Cristia
Journal:  Open Mind (Camb)       Date:  2021-11-25

7.  White matter abnormalities across different epilepsy syndromes in adults: an ENIGMA-Epilepsy study.

Authors:  Sean N Hatton; Khoa H Huynh; Leonardo Bonilha; Eugenio Abela; Saud Alhusaini; Andre Altmann; Marina K M Alvim; Akshara R Balachandra; Emanuele Bartolini; Benjamin Bender; Neda Bernasconi; Andrea Bernasconi; Boris Bernhardt; Núria Bargallo; Benoit Caldairou; Maria E Caligiuri; Sarah J A Carr; Gianpiero L Cavalleri; Fernando Cendes; Luis Concha; Esmaeil Davoodi-Bojd; Patricia M Desmond; Orrin Devinsky; Colin P Doherty; Martin Domin; John S Duncan; Niels K Focke; Sonya F Foley; Antonio Gambardella; Ezequiel Gleichgerrcht; Renzo Guerrini; Khalid Hamandi; Akari Ishikawa; Simon S Keller; Peter V Kochunov; Raviteja Kotikalapudi; Barbara A K Kreilkamp; Patrick Kwan; Angelo Labate; Soenke Langner; Matteo Lenge; Min Liu; Elaine Lui; Pascal Martin; Mario Mascalchi; José C V Moreira; Marcia E Morita-Sherman; Terence J O'Brien; Heath R Pardoe; José C Pariente; Letícia F Ribeiro; Mark P Richardson; Cristiane S Rocha; Raúl Rodríguez-Cruces; Felix Rosenow; Mariasavina Severino; Benjamin Sinclair; Hamid Soltanian-Zadeh; Pasquale Striano; Peter N Taylor; Rhys H Thomas; Domenico Tortora; Dennis Velakoulis; Annamaria Vezzani; Lucy Vivash; Felix von Podewils; Sjoerd B Vos; Bernd Weber; Gavin P Winston; Clarissa L Yasuda; Alyssa H Zhu; Paul M Thompson; Christopher D Whelan; Neda Jahanshad; Sanjay M Sisodiya; Carrie R McDonald
Journal:  Brain       Date:  2020-08-01       Impact factor: 13.501

Review 8.  An overview of the first 5 years of the ENIGMA obsessive-compulsive disorder working group: The power of worldwide collaboration.

Authors:  Odile A van den Heuvel; Premika S W Boedhoe; Sara Bertolin; Willem B Bruin; Clyde Francks; Iliyan Ivanov; Neda Jahanshad; Xiang-Zhen Kong; Jun Soo Kwon; Joseph O'Neill; Tomas Paus; Yash Patel; Fabrizio Piras; Lianne Schmaal; Carles Soriano-Mas; Gianfranco Spalletta; Guido A van Wingen; Je-Yeon Yun; Chris Vriend; H Blair Simpson; Daan van Rooij; Marcelo Q Hoexter; Martine Hoogman; Jan K Buitelaar; Paul Arnold; Jan C Beucke; Francesco Benedetti; Irene Bollettini; Anushree Bose; Brian P Brennan; Alessandro S De Nadai; Kate Fitzgerald; Patricia Gruner; Edna Grünblatt; Yoshiyuki Hirano; Chaim Huyser; Anthony James; Kathrin Koch; Gerd Kvale; Luisa Lazaro; Christine Lochner; Rachel Marsh; David Mataix-Cols; Pedro Morgado; Takashi Nakamae; Tomohiro Nakao; Janardhanan C Narayanaswamy; Erika Nurmi; Christopher Pittenger; Y C Janardhan Reddy; João R Sato; Noam Soreni; S Evelyn Stewart; Stephan F Taylor; David Tolin; Sophia I Thomopoulos; Dick J Veltman; Ganesan Venkatasubramanian; Susanne Walitza; Zhen Wang; Paul M Thompson; Dan J Stein
Journal:  Hum Brain Mapp       Date:  2020-03-10       Impact factor: 5.399

9.  Subcortical Brain Volume, Regional Cortical Thickness, and Cortical Surface Area Across Disorders: Findings From the ENIGMA ADHD, ASD, and OCD Working Groups.

Authors:  Premika S W Boedhoe; Daan van Rooij; Martine Hoogman; Jos W R Twisk; Lianne Schmaal; Yoshinari Abe; Pino Alonso; Stephanie H Ameis; Anatoly Anikin; Alan Anticevic; Celso Arango; Paul D Arnold; Philip Asherson; Francesca Assogna; Guillaume Auzias; Tobias Banaschewski; Alexander Baranov; Marcelo C Batistuzzo; Sarah Baumeister; Ramona Baur-Streubel; Marlene Behrmann; Mark A Bellgrove; Francesco Benedetti; Jan C Beucke; Joseph Biederman; Irene Bollettini; Anushree Bose; Janita Bralten; Ivanei E Bramati; Daniel Brandeis; Silvia Brem; Brian P Brennan; Geraldo F Busatto; Sara Calderoni; Anna Calvo; Rosa Calvo; Francisco X Castellanos; Mara Cercignani; Tiffany M Chaim-Avancini; Kaylita C Chantiluke; Yuqi Cheng; Kang Ik K Cho; Anastasia Christakou; David Coghill; Annette Conzelmann; Ana I Cubillo; Anders M Dale; Sara Dallaspezia; Eileen Daly; Damiaan Denys; Christine Deruelle; Adriana Di Martino; Ilan Dinstein; Alysa E Doyle; Sarah Durston; Eric A Earl; Christine Ecker; Stefan Ehrlich; Benjamin A Ely; Jeffrey N Epstein; Thomas Ethofer; Damien A Fair; Andreas J Fallgatter; Stephen V Faraone; Jennifer Fedor; Xin Feng; Jamie D Feusner; Jackie Fitzgerald; Kate D Fitzgerald; Jean-Paul Fouche; Christine M Freitag; Egill A Fridgeirsson; Thomas Frodl; Matt C Gabel; Louise Gallagher; Tinatin Gogberashvili; Ilaria Gori; Patricia Gruner; Deniz A Gürsel; Shlomi Haar; Jan Haavik; Geoffrey B Hall; Neil A Harrison; Catharina A Hartman; Dirk J Heslenfeld; Yoshiyuki Hirano; Pieter J Hoekstra; Marcelo Q Hoexter; Sarah Hohmann; Marie F Høvik; Hao Hu; Chaim Huyser; Neda Jahanshad; Maria Jalbrzikowski; Anthony James; Joost Janssen; Fern Jaspers-Fayer; Terry L Jernigan; Dmitry Kapilushniy; Bernd Kardatzki; Georgii Karkashadze; Norbert Kathmann; Christian Kaufmann; Clare Kelly; Sabin Khadka; Joseph A King; Kathrin Koch; Gregor Kohls; Kerstin Konrad; Masaru Kuno; Jonna Kuntsi; Gerd Kvale; Jun Soo Kwon; Luisa Lázaro; Sara Lera-Miguel; Klaus-Peter Lesch; Liesbeth Hoekstra; Yanni Liu; Christine Lochner; Mario R Louza; Beatriz Luna; Astri J Lundervold; Charles B Malpas; Paulo Marques; Rachel Marsh; Ignacio Martínez-Zalacaín; David Mataix-Cols; Paulo Mattos; Hazel McCarthy; Jane McGrath; Mitul A Mehta; José M Menchón; Maarten Mennes; Mauricio Moller Martinho; Pedro S Moreira; Astrid Morer; Pedro Morgado; Filippo Muratori; Clodagh M Murphy; Declan G M Murphy; Akiko Nakagawa; Takashi Nakamae; Tomohiro Nakao; Leyla Namazova-Baranova; Janardhanan C Narayanaswamy; Rosa Nicolau; Joel T Nigg; Stephanie E Novotny; Erika L Nurmi; Eileen Oberwelland Weiss; Ruth L O'Gorman Tuura; Kirsten O'Hearn; Joseph O'Neill; Jaap Oosterlaan; Bob Oranje; Yannis Paloyelis; Mara Parellada; Paul Pauli; Chris Perriello; John Piacentini; Fabrizio Piras; Federica Piras; Kerstin J Plessen; Olga Puig; J Antoni Ramos-Quiroga; Y C Janardhan Reddy; Andreas Reif; Liesbeth Reneman; Alessandra Retico; Pedro G P Rosa; Katya Rubia; Oana Georgiana Rus; Yuki Sakai; Anouk Schrantee; Lena Schwarz; Lizanne J S Schweren; Jochen Seitz; Philip Shaw; Devon Shook; Tim J Silk; H Blair Simpson; Norbert Skokauskas; Juan Carlos Soliva Vila; Anastasia Solovieva; Noam Soreni; Carles Soriano-Mas; Gianfranco Spalletta; Emily R Stern; Michael C Stevens; S Evelyn Stewart; Gustavo Sudre; Philip R Szeszko; Leanne Tamm; Margot J Taylor; David F Tolin; Michela Tosetti; Fernanda Tovar-Moll; Aki Tsuchiyagaito; Theo G M van Erp; Guido A van Wingen; Alasdair Vance; Ganesan Venkatasubramanian; Oscar Vilarroya; Yolanda Vives-Gilabert; Georg G von Polier; Susanne Walitza; Gregory L Wallace; Zhen Wang; Thomas Wolfers; Yuliya N Yoncheva; Je-Yeon Yun; Marcus V Zanetti; Fengfeng Zhou; Georg C Ziegler; Kathrin C Zierhut; Marcel P Zwiers; Paul M Thompson; Dan J Stein; Jan Buitelaar; Barbara Franke; Odile A van den Heuvel
Journal:  Am J Psychiatry       Date:  2020-06-16       Impact factor: 18.112

10.  In vivo hippocampal subfield volumes in bipolar disorder-A mega-analysis from The Enhancing Neuro Imaging Genetics through Meta-Analysis Bipolar Disorder Working Group.

Authors:  Unn K Haukvik; Tiril P Gurholt; Stener Nerland; Torbjørn Elvsåshagen; Theophilus N Akudjedu; Martin Alda; Dag Alnaes; Silvia Alonso-Lana; Jochen Bauer; Bernhard T Baune; Francesco Benedetti; Michael Berk; Francesco Bettella; Erlend Bøen; Caterina M Bonnín; Paolo Brambilla; Erick J Canales-Rodríguez; Dara M Cannon; Xavier Caseras; Orwa Dandash; Udo Dannlowski; Giuseppe Delvecchio; Ana M Díaz-Zuluaga; Theo G M van Erp; Mar Fatjó-Vilas; Sonya F Foley; Katharina Förster; Janice M Fullerton; José M Goikolea; Dominik Grotegerd; Oliver Gruber; Bartholomeus C M Haarman; Beathe Haatveit; Tomas Hajek; Brian Hallahan; Mathew Harris; Emma L Hawkins; Fleur M Howells; Carina Hülsmann; Neda Jahanshad; Kjetil N Jørgensen; Tilo Kircher; Bernd Krämer; Axel Krug; Rayus Kuplicki; Trine V Lagerberg; Thomas M Lancaster; Rhoshel K Lenroot; Vera Lonning; Carlos López-Jaramillo; Ulrik F Malt; Colm McDonald; Andrew M McIntosh; Genevieve McPhilemy; Dennis van der Meer; Ingrid Melle; Elisa M T Melloni; Philip B Mitchell; Leila Nabulsi; Igor Nenadić; Viola Oertel; Lucio Oldani; Nils Opel; Maria C G Otaduy; Bronwyn J Overs; Julian A Pineda-Zapata; Edith Pomarol-Clotet; Joaquim Radua; Lisa Rauer; Ronny Redlich; Jonathan Repple; Maria M Rive; Gloria Roberts; Henricus G Ruhe; Lauren E Salminen; Raymond Salvador; Salvador Sarró; Jonathan Savitz; Aart H Schene; Kang Sim; Marcio G Soeiro-de-Souza; Michael Stäblein; Dan J Stein; Frederike Stein; Christian K Tamnes; Henk S Temmingh; Sophia I Thomopoulos; Dick J Veltman; Eduard Vieta; Lena Waltemate; Lars T Westlye; Heather C Whalley; Philipp G Sämann; Paul M Thompson; Christopher R K Ching; Ole A Andreassen; Ingrid Agartz
Journal:  Hum Brain Mapp       Date:  2020-10-19       Impact factor: 5.399

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.