Literature DB >> 31010934

Structured Genome-Wide Association Studies with Bayesian Hierarchical Variable Selection.

Yize Zhao1, Hongtu Zhu2, Zhaohua Lu3, Rebecca C Knickmeyer4, Fei Zou5.   

Abstract

It becomes increasingly important in using genome-wide association studies (GWAS) to select important genetic information associated with qualitative or quantitative traits. Currently, the discovery of biological association among SNPs motivates various strategies to construct SNP-sets along the genome and to incorporate such set information into selection procedure for a higher selection power, while facilitating more biologically meaningful results. The aim of this paper is to propose a novel Bayesian framework for hierarchical variable selection at both SNP-set (group) level and SNP (within group) level. We overcome a key limitation of existing posterior updating scheme in most Bayesian variable selection methods by proposing a novel sampling scheme to explicitly accommodate the ultrahigh-dimensionality of genetic data. Specifically, by constructing an auxiliary variable selection model under SNP-set level, the new procedure utilizes the posterior samples of the auxiliary model to subsequently guide the posterior inference for the targeted hierarchical selection model. We apply the proposed method to a variety of simulation studies and show that our method is computationally efficient and achieves substantially better performance than competing approaches in both SNP-set and SNP selection. Applying the method to the Alzheimers Disease Neuroimaging Initiative (ADNI) data, we identify biologically meaningful genetic factors under several neuroimaging volumetric phenotypes. Our method is general and readily to be applied to a wide range of biomedical studies.
Copyright © 2019 by the Genetics Society of America.

Entities:  

Keywords:  Bayesian variable selection; Markov chain Monte Carlo; SNP-set; genome-wide association studies; imaging genetics

Mesh:

Substances:

Year:  2019        PMID: 31010934      PMCID: PMC6553832          DOI: 10.1534/genetics.119.301906

Source DB:  PubMed          Journal:  Genetics        ISSN: 0016-6731            Impact factor:   4.562


IN modern genetics, genome-wide association studies (GWAS) has become a popular tool to study complex human diseases (Walsh ; Wang ; Hibar ). The goal of GWAS is to identify single nucleotide polymorphisms (SNPs) associated with complex traits. Over the last few years, improvement in genotyping technology has enriched the measurements of SNPs to >1 million (Altshuler ). Due to the ultrahigh-dimensionality of SNPs, most GWAS approaches analyze one SNP at a time to test marginal association of the SNP with phenotype. However, in many scenarios, large differences can exist between marginal effects of SNPs and their joint effects (He and Lin 2011). Thus, it is imperative to carry out whole-genome GWAS that considers all SNPs together. There are at least two challenges associated with whole-genome GWAS. The first one is the “large p small n” problem. To address this issue, various regularization or screening methods (Tibshirani 1996; Fan and Li 2001; Efron ; Zou and Hastie 2005; Zou 2006; Fan and Lv 2008) were proposed and recently extended to the context of GWAS (Hoggart ; Wu ; Cho ; He and Lin 2011; Sampson ; Jiang ; Bao and Wang 2017; Huang ). As an alternative, Bayesian methods also play a prominent role in solving variable selection problem. O’Hara and Sillanpää (2009) provided an overview of several commonly used Bayesian variable selection methods and posterior simulation algorithms, such as the Gibbs variable selection (GVS) (Dellaportas ) and the stochastic search variable selection (SSVS) (George and McCulloch 1993). Compared with regularization methods, Bayesian models have the natural advantage to quantify uncertainty and combine prior information. Their recent application on GWAS also showed a higher detection power by simultaneously fitting multiple marker effects and implicitly correcting biological structures (Sahana ; Dashab ; Kärkkäinen and Sillanpää 2012). One major limitation to apply Bayesian variable selection models on GWAS is the intensive computation. Therefore, existing approaches made attempts to explore different inference mechanisms to reduce the computational cost or improve the mixing of Markov chain. For instance, besides traditional Markov chain Monte Carlo (MCMC) algorithm (Guan ), variational Bayes (Carbonetto ), evolutionary stochastic search (Bottolo ) and other variations of stochastic searching algorithms (Briollais ; Yang ) were developed under Bayesian sparse models for multi-SNP analysis. Alternatively, Bayesian lasso (Li ; Jiang ) and Bayesian mixed model (Zhou ; Zhou 2014) were also considered to improve the scalability in the presence of ultrahigh-dimensional SNP data. Another challenge is caused by biological architecture. Multiple causal SNPs may be located in a single region, each with a small effect. In order to increase power to map them, it is desirable to consider them simultaneously and perform SNP-set analysis. A variety of methods were proposed for SNP-set analysis, including, but not limited to, Kwee , Wang and Abbott (2008), Wu and a more recent Bayesian latent sparse model by Lu . Tzeng provided an overview of different marker-set approaches for gene-trait association and appealing features of set-based methods compared with those using individual SNPs. For the marker-set methods, grouping strategies to define SNP-sets play a vital role in practice. As suggested by Wu , the ones incorporating biological information, for example grouping SNPs by genes, pathways or haplotype/linkage disequilibrium (LD) blocks, tend to gain more power. Based on the defined SNP-sets, analysis could be carried out through weighted sum of genotypes (Wang and Elston 2007; Price ), U-statistics (Tzeng ; Wei ), or variance-component methods (Tzeng and Zhang 2007; Wu ). Meanwhile, the use of SNP-sets can also reduce the number of predictors as well as alleviate the collinearity issue since the correlations are much smaller among SNP-sets than SNPs. The appealing features of SNP-set analysis motivate us to incorporate set-wise information into the selection procedure. In this paper, we develop a Bayesian hierarchical variable selection model to carry out whole-genome association analysis and achieve SNP/SNP-set trait association mapping. Our variable selection approach is inherently hierarchical, and involves selection at both SNP-set level and individual SNP level. Although there is a broad literature on Bayesian variable selection under high or ultrahigh-dimensional feature space (Bottolo et al. 2010; Johnson and Rossell 2012; Johnson 2013), few efficient hierarchical variable selection methods have been developed. Stingo considered gene expression data, and adopted Bayesian spike-and-slab priors to simultaneously select genes and pathways. Rockova et al. (2014) proposed a two-step procedure to carry out hierarchical variable selection under Bayesian group Lasso. Combining spike-and-slab priors with shrinkage priors, Duan and Thomas (2013), Zhang ,b), Liquet proposed similar modeling frameworks, and achieved group level selection and within-group shrinkage. Though the existing approaches are promising, they suffered with two limitations. First, the methods employ traditional MCMC algorithm, which are computational intensive and difficult to scale up. Second, the group level selection does not utilize any structural information, which could lead to poor performance. Recently, Tang proposed a Bayesian hierarchical generalized linear models incorporating group information, and developed an EM algorithm for parameter estimation. However, the method does not impose group-level sparsity and may be less powerful under sparse signal like GWAS application. In this paper, we develop a Bayesian hierarchical selection model, named Sparse Group Hierarchical Sampling (SGHS), with an efficient posterior inference algorithm. The key idea of SGHS is to reallocate posterior computation spending on the potential signal and noise parts via a “smart” proposal distribution. The sampling scheme for the proposal distribution is specified by an auxiliary model constructed under set-wise variable selection using factor regression. Such a modeling scheme allows the MCMC algorithm to explore the entire sample space more efficiently and dramatically mitigate the computational burden of updating large-scale unknown parameters. Simultaneously, grouping and structural information are integrated within the posterior inference, leading to the improvement on selection accuracy and results interpretability. The remainder of this article is organized as follows. In Model specification, we present our basic model for variable selection, prior specifications, and a standard posterior computation algorithm. In Sparse group hierarchical sampling, we propose our SGHS for hierarchical variable selection. We conduct simulation studies in Simulation studies to assess the performance of our proposed approach, and in The Alzheimer’s Disease neuroimaging initiative, we apply the method to the ADNI data to identify disease related genetic information. Finally, we conclude with a Discussion.

Materials and Methods

Model specification

We introduce the variable selection model in a Bayesian framework. Our description is based on GWAS, but the proposed method is general and readily extended to other applications. Assume there are n subjects in the data. For subject , is a trait of interest, such as the brain volume of a region of interest (ROI), and is a -vector of clinical variables, including an intercept term. Suppose that the whole genome contains J SNPs, based on which, K SNP-sets are defined with SNPs included in SNP-set k. We assume the SNP-sets are mutually exclusive, then the total number of SNPs . We let denote the genotype of SNP j within SNP-set k representing by the minor alleles, and we also conduct a subsequent normalization for each SNP. We consider a standard regression model for hierarchical variable selection given bywith the residual error . Here is the set-wise selection indicator describing the selection status for SNP-set k, and is the individual one for SNP j within SNP-set k. Moreover, and are regression coefficients with . Our goal is to identify risk SNP-sets along with specific SNPs by using the selection indicators . Particularly, is included in model (1) if both and are nonzero. We can further write model (1) in a more compact formwhere and . Here “” is the individual-to-set mapping matrix with “” representing the Hadamard product (Styan 1973). As is the general case in practice, we assume that “signals” are sparse with the cardinality of the active set . It is worth noting that, in practice, SNP-sets can be overlapped based on certain biologically grouping strategy, e.g., genes, pathways. In such cases, we can randomly assign a SNP to one of its overlapped SNP-sets. We introduce priors for all the parameters in model (1). Specifically, we assign conjugate Gaussian priors to the regression coefficients asand conjugate hyper-priors for the variance parameters and . An equivalent model formulation is to define a coefficient , which results in the well-known point mass mixture prior with a point mass at zero. Compared with mixture prior, the above specification enables a more efficient updated scheme for coefficients, which we will explain later. In terms of the selection indicators and , we assume separate independent Bernoulli priorswhere and with controlling the proportion of SNP-sets in the model, and determining the proportion of significant SNPs. We finally assign a prior for the variance of residual error . Parameters included in the posterior inference are , , , , , and η with the joint conditional posterior distributionStandard MCMC algorithm can be implemented to update each parameter from its full conditional distribution. The details of updating scheme for all the parameters are provided in Appendix A1. As noted, to update , we resort to a block updating scheme by dividing the whole long vector into two blocks, and corresponding to the selected () and unselected () predictors. Thus, the conditional distributions of and are given bywhere and with corresponding to the active set. Such vector-wise updating scheme can dramatically reduce computational cost and lead to better mixing than the single-site Gibbs sampling.

Sparse group hierarchical sampling

Standard MCMC algorithm becomes inefficient in the presence of high-dimensional data, and even computationally infeasible under GWAS with tens of thousands of predictors. Therefore, in this section, we propose the SGHS scheme to overcome the computational complexity in the posterior computation. The key idea of SGHS is to construct an auxiliary model based on set-wise variable selection, which subsequently realizes a reweight of posterior computational effort on potential signals and noises in the target model to efficiently search among model spaces. Specifically, we first introduce an auxiliary set-wise indicator and define each of their elementsThrough this definition, the auxiliary set-wise indicator is completely determined by the SNP-wise selection indicator, by the selection consistency between two levels, and the structure of stays the same as that of the set-wise selection indicator . Accordingly, the posterior distribution for the parameters becomeswhere only if (7), and zero otherwise. In the right-hand side of (8), the first term consistent with Equation 5 is the main probability part, while the second term captures the connection of the selection information between SNP-set level and SNP level, inducing the incorporation of group membership. A comparison between (5) and (8) shows that the posterior probabilities and sampling schemes for parameters , , η, and hyper-parameters , keep consistent with those in the standard MCMC algorithm. As for the selection indicators and and auxiliary indicator , we take a novel approach to jointly update via a Metropolis-Hastings (M-H) step by constructing a proposal distributionwhere the subscripts “c” and “*” denote the current value and the proposed value, and “” represents all the other parameters. Here specifies the sampling scheme for the auxiliary indicator , and we design it to depend only on the data as shown in Equation 9. Such specification removes the interference of other parameters, which allows function to be achieved by a separate model with variable selection only at SNP-set level. We refer to this model as an auxiliary model (distinguished from the target model) with the goal to help the posterior simulation of the target model. Given the sampled value , the sampling scheme for is specified by function which induces the incorporation of information from SNP-set level selection. Under such a proposal distribution, function is a receiver with the set-wise selection information “copied” from the auxiliary model, and function is a transmitter to pass such information from to the target model. In the following sections, we will discuss the choice of and that could lead to a more efficient posterior sample procedure.

Receiver function :

As discussed before, the sampling procedure of receiver function is induced by an auxiliary SNP-set level selection model. In other words, realization of can be directly simulated from the posterior distribution of the auxiliary model. As a natural choice of an auxiliary model, following model (1), we havewhere . We use a superscript “a” in (10) to distinguish the corresponding parameters in the auxiliary model from the target model. Model (10) turns out to be less attractive in general cases where not all the SNPs within a SNP-set are predictive for phenotype of interest. Since the assumption of nonzero coefficients in the risk SNP-sets is conflicted with the within-group level sparsity, it is difficult for model (10) to approximate the truth. To address this issue, we resort to an alternative set-wise selection model based on an empirical factor (principal component) regression under each SNP-set. Specifically, we adopt a reduced rank singular-value decomposition (SVD) on each , namely , for . Here, with is an factor matrix subject to , where is a diagonal matrix formed by positive singular values of , and is the loadings matrix subject to . By replacing each with the reduced SVD representation, the new auxiliary model becomeswhere is the coefficient for factor l within SNP-set k. In GWAS data where there are strong correlations among predictors, to avoid overfitting, we could also allow certain truncation by specifying a cutoff on the number of factors included with . Comparing with , under model (11), we realize a dimension reduction of the predictors from J to L with , where . For selected SNP sets, the overall nonzero assumption of vector also fits many scenarios in real applications. Thus, we use model (11) as the auxiliary model in our applications. Again, the choice of an auxiliary model is not uniquely determined since it serves only as a platform to influence posterior sampling of the main model, but model (11), for example, has been shown to perform well in a wide range of scenarios. Following the prior specification in (3) for the target model, we assign similar conjugate priors for the coefficientsand for the variance parameters , and . We assign independent Bernoulli priors for the auxiliary indicatorswhere and with controls the proportion of selection. We further introduce a hyper-prior for each proportion parameter as follows:with the shape parameters. The hyper-prior (14) distinguishes the posterior probabilities between risk and nonrisk SNP-sets via a mixture of a zero point mass and a spread Beta distribution. By combining priors (13) and (14), we can integrate out which results in independent Bernoulli priors for indicator with proportion parameter . In this case, we use (13) and (14) to ensure sparsity in the group level auxiliary model, which subsequently brings impact to the target model.

Transmitter function :

To efficiently sample from the large feature space, we propose as follows:withwhere ϕ, and are tuning parameters to make sure detailed balance in the M-H step is satisfied, as well as to allow enough information to be borrowed from proposals. Specifically, ϕ determines the amount of difference between SNP-set selection and the proposal, and if a SNP-set is currently unselected, controls the sparsity for the selection of its SNPs; otherwise, influences the agreement of the SNP-level selection between proposed and current status. In practice, we specify , and to allow an efficient transmission of the selection information. To implement this M-H step, we first draw with simulated by the posterior distribution of in the auxiliary model. Then, we draw . Finally, we calculateand set when with . To further improve the mixing of Markov chains, in addition to the M-H step, we also conduct a further moving step for under from its full conditional (32). Since the true signal is sparse, such a moving step does not require a heavy computation. Under the specification of this proposal distribution, in our SGHS scheme, each SNP has a positive probability to be selected or unselected. However, instead of updating all the SNPs at each iteration of the posterior simulation, a large amount of nonrisk variables (noises) have been directly “labeled” iteration by iteration (without updating), which allows us to spend most of the computation updating potential signal part. As a result, the computational efficiency is dramatically improved compared with the existing MCMC algorithm. A detailed MCMC algorithm for the SGHS scheme is provided in Appendix A2.

Data availability

We used ADNI1 genetics and MRI image data that are available through application (http://adni.loni.usc.edu/about/adni1/). All the ADNI data are shared without embargo through the LONI Image and Data Archive (IDA).

Results

Simulation studies

We conduct simulation studies to evaluate the finite-sample performance of SGHS. Our goal is to select genetic markers that are highly associated with outcome of interest. We focus on the comparison between the proposed SGHS and existing methods include lasso (Lasso) (Tibshirani 1996), smoothly slipped absolute deviation (SCAD) (Fan and Li 2001), sparse-group Lasso (SGL) (Friedman ), functional genome-wide association analysis (FGWAS) (Huang ), Bayesian variable selection with posterior inference via model averaging and subset selection (piMASS) (Guan ), and genome-wide efficient mixed model association (GEMMA) based on Bayesian sparse linear mixed model (Zhou 2014). To better assess each method under GWAS, all the simulation scenarios are designed to mimic real genetic data. We generate subjects with the genetic information simulated from the Hapmap projects 2009-02 phaseIII data (International HapMap 3 Consortium 2010). Specifically, for each subject, we randomly combine two haplotypes from the CEPH population to form its genotypes. We consider both a low dimensional scenario by randomly selecting 5000 SNPs, and a high-dimensional one with 100,000 SNPs. Under each scenario, we determine SNP-sets (LD blocks) by starting from an initial SNP m with a putative block of SNPs and considering sub-blocks with until >50% of elements in the corresponding matrix of value surpass the threshold κ. To further assess the impact of LD structure, we set different thresholds with and 0.1 to construct the grouping information, based on which, we consider the following two cases of signal patterns to evaluate the robustness of variable selection. Case 1: We randomly select 10 risk SNP-sets. Within each of them, we randomly set 10 SNPs as risk ones. Case 2: We randomly select 10 risk SNP-sets. Within SNP-set k with , we randomly set k SNPs as risk ones; and within SNP-sets with , all the SNPs are risk ones. We denote as the number of risk SNPs in the SNP-set k. Based on the two signal patterns, we consider a variety of settings for the value of nonzero regression coefficients in model (1), i.e., the regression coefficients of risk SNPs denoted by , as shown in Table 1. Specifically, in Case 1, where sparse signals exist, we consider six different settings. In Settings 1 and 2, as the starting point, we assign a unified genetic effect under different magnitudes. In Settings 3 and 4, the associations between the risk SNPs and the phenotype become more general but the majority of them are in the same sign. Finally, in Settings 5 and 6, we allow both positive and negative genetic effects within each SNP set to purposely create barriers for the internal SNP-set level selection. In Case 2, we further dilute the signals in the later half of risk SNP-sets while keep sparsity in the former half in Settings I and II, and our design guarantees the average strength of signals in each risk SNP-set is comparable. For each setting, we generate 100 Monte Carlo (MC) datasets to assess feature selection performance among all the methods.
Table 1

Simulation design: different settings of nonzero coefficients under two cases

Setting
Case 11. β1=0.5
2. β1=1
3. β1N(1,0.5I)
4. β1N(3,0.5I)
5. β1=(1,1,1,1,)
6. β1N(0,0.5I)
Case 2I. βjkN(5/dk,0.5I)
II. βjkN(10/dk,0.5I)
To implement SGHS, we conduct posterior inference with random initials for 10,000 iterations with 5000 burn-in for both the auxiliary and target models. The average computational time per dataset is 3.4 min for low dimensional scenarios and 26.5 min for high dimensional ones (Matlab implementation, 3.4 GHz CPU, 8 GB Memory, Windows System) to finish the whole posterior inference and the convergence is checked by GR method (Gelman and Rubin 1992) as well as trace plots. To allow for a noninformative hyper-prior, we assign a relative large value for and as 10. We also set to accommodate sparse signals in the auxiliary model, and less informative and in the target model. Finally, we truncate the number of factors in each subset by looking for the minimum number of singular vectors that can explain ∼ of the total variance. For the competing methods, we use R packages glmnet, ncvreg, and SGL to implement Lasso, SCAD, and SGL, and public released pipelines for FGWAS, GEMMA, and piMASS. For the Bayesian algorithms GEMMA and piMASS, we set all the tuning parameters as recommended by the manuals, and the average computational time is 10.9/48.6 min for GEMMA and 2.1/16.6 min for piMASS under low/high dimensional scenarios. Finally, the feature selection performance is assessed by sensitivity (Sens); specificity (Spec); Youden’s J statistic (J Stat), which equals sensitivity specificity; and area under the receiver operating characteristic curve (AUC) for all the methods. The simulation results are summarized in Table 2 and Table 3 under low and high dimensions. To determine selection status for the Bayesian methods, i.e., GEMMA, piMASS, and SGHS, we use 0.5 as cutoff on the marginal posterior probability of selection indicator (Barbieri and Berger 2004). We first compare our method to the competing approaches. For all the settings under different LD thresholds κ and dimensions, our proposed method outperforms all competing methods in selection accuracy with a higher sensitivity, J Stat and AUC. Specifically, Settings 1 and 2 are constructed with a unified genetic effect among risk SNPs, and SGHS achieves the highest AUC and a much higher J Stat than its competitors under different dimensions and κ. In Settings 3 and 4, with a more general genetic effect, SGHS maintains its satisfactory performance in feature selection. In Settings 5 and 6, where positively and negatively associated SNPs within a causal SNP-set are almost equally distributed, while we notice decreases on AUC and J Stat for almost all the methods, SGHS still achieves a superior selection performance compared to other methods along with a satisfactory AUC. This reflects, to some extent, the robustness of SGHS to the extreme settings. Finally, the diluted signal patterns in Settings I and II further deteriorates the performance for LASSO, SCAD, SGL, piMASS, and GEMMA, but brings little impact on FGWAS and SGHS, with the latter remaining the best performer. Overall, Table 2 and Table 3 show a general pattern that SGHS works considerably better on detecting true risk features, as indicated by much higher sensitivities. Even though false positives have also been brought in as a consequence, we still obtain a remarkable feature selection accuracy.
Table 2

Simulation results: feature selection performance with

κ=0.01κ=0.05κ=0.1
SettingMethodSensSpecJ StatAUCSensSpecJ StatAUCSensSpecJ StatAUC
1Lasso0.5000.9220.4220.7960.5010.9320.4330.8100.5150.9390.4540.803
SCAD0.1680.9630.1310.6980.1610.9680.1290.7170.1670.9700.1370.695
SGL0.6840.9040.5880.8150.6880.9260.6140.8170.6480.9360.5840.798
FGWAS0.4440.9570.4010.7480.5410.9500.4910.7730.5990.9510.5500.801
piMASS0.1280.9980.1260.8160.1460.9980.1440.8160.1120.9980.1100.839
GEMMA0.1780.9930.1710.7210.1760.9950.1710.7170.1440.9960.1400.727
SGHS0.8060.9190.7250.9570.8090.9450.7540.9660.8030.9510.7540.974
2Lasso0.5440.9120.4560.8110.5880.9280.5160.8290.5730.9330.5060.816
SCAD0.1710.9660.1370.7130.1600.9700.1300.7170.1820.9700.1520.703
SGL0.6880.9100.5980.8140.6880.9270.5150.8150.6440.9350.5790.802
FGWAS0.4430.9410.3840.7510.5530.9490.5020.7790.6060.9510.5570.797
piMASS0.1960.9970.1930.8130.1960.9980.1940.9820.1600.9980.1580.840
GEMMA0.2980.9780.2760.7010.2840.9790.2630.6890.2960.9780.2740.722
SGHS0.8500.9150.7650.9470.8530.9410.7940.9740.8610.9490.8100.977
3Lasso0.5220.9180.4400.8000.5550.9330.4880.8130.5970.9340.5310.815
SCAD0.1750.9660.1410.7100.1640.9700.1340.6980.1750.9710.1460.697
SGL0.6530.9150.5680.7950.6610.9290.5900.8060.6280.9380.5660.789
FGWAS0.4680.9460.4140.6970.5660.9490.5150.6960.5790.9520.5310.797
piMASS0.2160.9980.2140.8420.1820.9980.1800.8390.1480.9980.1460.841
GEMMA0.2620.9770.2390.6710.2740.9790.2530.6770.2400.9810.2210.656
SGHS0.8210.9130.7340.9520.8110.9450.7560.9620.8030.9490.7520.968
4Lasso0.6030.9240.5270.8220.6010.9420.5430.8400.5930.9470.5400.827
SCAD0.1700.9670.1370.7140.1610.9700.1310.7130.1650.9710.1360.690
SGL0.6880.9110.5990.8150.6770.9270.6040.8100.6520.9370.5890.805
FGWAS0.4300.9530.3830.7550.5500.9500.5000.7270.5990.9510.5000.805
piMASS0.3260.9930.3190.8180.2740.9940.2680.8150.2840.9950.2790.821
GEMMA0.4020.9530.3550.6980.3500.9620.3120.6820.4420.9600.4020.716
SGHS0.8270.9190.7460.9640.8470.9370.7840.9740.8280.9470.7750.975
5Lasso0.4650.9160.3810.4890.4230.9260.3490.4750.4240.9250.3490.503
SCAD0.1790.9660.1450.5040.1850.9690.1540.4820.1700.9690.1490.499
SGL0.5230.9090.4320.5120.4480.9350.3830.4980.4950.9380.4330.502
FGWAS0.4310.9580.5090.5260.4430.9640.4070.6220.4900.9630.4530.719
piMASS0.1820.9970.1790.7650.1620.9980.1600.7670.1540.9980.1520.709
GEMMA0.1780.9920.1700.6520.2240.9950.2190.6840.1720.9960.1680.662
SGHS0.7950.9230.7180.9520.7840.9510.7350.9610.7820.9570.7390.957
6Lasso0.3900.9390.3290.4500.3620.9490.3110.4940.3760.9480.3240.505
SCAD0.1860.9590.1450.4900.1620.9650.1270.4870.1680.9650.1330.499
SGL0.5540.8940.4480.5020.5420.9180.4600.5040.5390.9260.4650.507
FGWAS0.4330.9560.5090.5280.4360.9630.3990.5750.4460.9660.4120.621
piMASS0.1000.9990.0990.8970.1040.9990.1030.8040.0800.9990.0790.805
GEMMA0.1240.9970.1230.7280.1200.9980.1180.7500.0920.9980.0900.770
SGHS0.6600.9600.6200.9270.6460.9630.6090.9420.7140.9560.6600.965
ILasso0.4280.9020.3300.7330.4450.9270.4720.7490.4560.9300.3860.743
SCAD0.0980.9130.0110.6600.0930.9360.0290.6400.0910.9420.0330.637
SGL0.4190.9200.3390.6580.4350.9350.3700.6860.4970.9360.4330.697
FGWAS0.3870.9940.3810.8120.6020.9930.5950.8270.6400.9920.6320.836
piMASS0.1160.9990.1150.8310.1280.9990.1270.8350.1410.9990.1400.847
GEMMA0.2180.9890.2070.6590.2400.9900.2300.6640.2570.9870.2440.666
SGHS0.7820.9650.7470.9770.7720.9720.7440.9820.7870.9780.7650.981
IILasso0.4430.9040.3470.7440.4690.9310.4000.7650.4650.9320.3970.755
SCAD0.0990.9130.0110.6590.1080.9370.0450.6560.0910.9420.0330.640
SGL0.4310.9190.3500.6700.4480.9350.3830.6760.4790.9370.4160.700
FGWAS0.3880.9930.3810.8080.6020.9930.5950.8220.5710.9930.5640.756
piMASS0.1470.9990.1460.8250.1680.9980.1660.8400.1650.9990.1640.831
GEMMA0.2140.9860.2000.6280.2640.9860.2500.6560.2730.9890.2620.656
SGHS0.8780.9510.8290.9520.8460.9350.7810.9720.8360.9600.7960.981

Sens, the average sensitivity; Spec, the average specificity; J Stat, the average Youden’s J statistic; and AUC, the average area under the receiver operating characteristic curve.

Table 3

Simulation results: feature selection performance with

κ=0.01κ=0.05κ=0.1
SettingMethodSensSpecJ StatAUCSensSpecJ StatAUCSensSpecJ StatAUC
1Lasso0.4690.9970.4660.8160.450.9970.4470.7930.4660.9970.4630.797
SCAD0.1550.9970.1520.7380.1610.9980.1590.7120.1470.9980.1450.709
SGL0.7380.9930.7310.8530.6740.9950.6690.8210.6700.9970.6670.832
FGWAS0.4580.9980.4560.5240.5160.9980.5140.7180.5490.9980.5470.801
piMASS0.0840.9990.0830.8240.1080.9990.1070.8250.1080.9990.1070.825
GEMMA0.0720.9990.0710.5520.0940.9990.0930.5640.0980.9990.0970.579
SGHS0.8580.9950.8530.9930.8860.9960.8820.9980.9040.9960.9000.999
2Lasso0.5460.9960.5420.8350.5520.9960.5480.8270.5530.9960.5490.829
SCAD0.1750.9980.1730.7380.1480.9990.1470.7120.1540.9990.1530.706
SGL0.7320.9960.7280.8570.6760.9970.6730.8210.6840.9970.6810.834
FGWAS0.4850.9980.4830.6490.5120.9980.5100.7860.5640.9980.5620.799
piMASS0.1380.9990.1370.8330.1680.9990.1670.8430.1680.9990.1670.843
GEMMA0.1280.9990.1270.5860.1520.9990.1510.6150.1820.9990.1810.633
SGHS0.9450.9830.9280.9960.9410.9850.9260.9970.9710.9770.9480.999
3Lasso0.5280.9960.5240.8110.5080.9970.5050.8070.5100.9960.5060.799
SCAD0.1750.9990.1740.7450.1530.9990.1520.6990.1510.9990.1500.689
SGL0.6960.9960.6920.8390.6580.9970.6550.7980.6740.9970.6710.829
FGWAS0.4460.9980.4440.5920.5200.9980.5180.6570.5320.9980.5300.796
piMASS0.1480.9990.1470.8520.1700.9990.1690.8470.1700.9990.1690.847
GEMMA0.1060.9990.1050.5690.1800.9990.1790.6120.1460.9990.1450.595
SGHS0.9240.9820.9060.9960.9290.9790.9080.9910.9690.9710.9400.995
4Lasso0.6060.9950.6010.8620.5960.9960.5920.8450.5960.9960.5920.839
SCAD0.1650.9980.1630.7290.1570.9990.1560.7150.1490.9990.1480.708
SGL0.7280.9960.7240.8530.6640.9970.6610.8130.6860.9970.6830.836
FGWAS0.4680.9980.4660.5800.5100.9980.5080.7870.5640.9980.5620.809
piMASS0.1860.9990.1850.8440.2420.9990.2410.8490.2420.9990.2410.849
GEMMA0.1260.9990.1250.5880.1960.9990.1950.6420.1960.9990.1950.626
SGHS0.9550.9500.9050.9880.9670.9490.9160.9900.9780.9540.9320.996
5Lasso0.3530.9970.3500.5050.3370.9970.3340.5010.3540.9970.3510.494
SCAD0.1820.9960.1780.4970.1650.9970.1620.5030.1670.9980.1650.484
SGL0.5780.9910.5690.4880.4900.9870.4770.4460.4640.9860.4500.551
FGWAS0.3720.9980.3700.3070.4420.9980.4400.4770.4650.9990.4640.409
piMASS0.1460.9990.1450.7720.1360.9990.1350.7910.1360.9990.1350.791
GEMMA0.1040.9990.1030.5620.1200.9990.1190.5740.1160.9990.1150.574
SGHS0.8750.9940.8690.9980.8660.9970.8630.9940.9080.9950.9030.998
6Lasso0.3450.9970.3420.4840.3200.9980.3180.5130.3180.9980.3160.506
SCAD0.1570.9970.1540.4940.1390.9970.1360.5010.1500.9970.1470.506
SGL0.5780.9870.5650.4860.5660.9930.5590.5420.5040.9920.4960.493
FGWAS0.3880.9980.3860.3970.4050.9980.4030.5520.4560.9980.4540.472
piMASS0.0640.9990.0630.8070.1080.9990.1070.8200.1080.9990.1070.820
GEMMA0.0680.9990.0670.5430.0980.9990.0970.5710.0900.9990.0890.554
SGHS0.7370.9970.7340.9980.7570.9970.7540.9830.8000.9970.7970.987
ILasso0.4360.9940.4300.7320.4440.9960.4400.7830.5000.9960.4960.802
SCAD0.0870.9960.0830.6380.0790.9970.0760.6850.0860.9970.0830.700
SGL0.4140.9960.4100.6930.4140.9970.4110.6920.4260.9970.4230.674
FGWAS0.5000.9990.4990.7720.5110.9990.5100.6830.6060.9990.6050.809
piMASS0.0730.9990.0720.8630.0890.9990.0880.8630.0890.9990.0880.863
GEMMA0.0870.9990.0860.5620.0800.9990.0790.5560.1310.9990.1300.588
SGHS0.9430.9510.8940.9900.9340.9550.8890.9840.9540.9580.9120.991
IILasso0.4720.9940.4660.7510.4750.9960.4710.7840.5100.9970.5070.818
SCAD0.0840.9960.0800.6540.0700.9970.0670.6710.0760.9970.0730.692
SGL0.4160.9960.4120.6900.4280.9970.4250.6880.4880.9970.4850.691
FGWAS0.4650.9990.4640.5660.5320.9990.5310.7850.6090.9990.6080.718
piMASS0.0800.9990.0790.8570.0660.9990.0650.8630.0660.9990.0650.863
GEMMA0.0730.9990.0720.5540.0820.9990.0810.5650.1130.9990.1120.578
SGHS0.9370.9340.8710.9850.9530.9440.8970.9900.9600.9530.9130.992

Sens, the average sensitivity; Spec, the average specificity; J Stat, the average Youden’s J statistic; and AUC, the average area under the receiver operating characteristic curve.

Sens, the average sensitivity; Spec, the average specificity; J Stat, the average Youden’s J statistic; and AUC, the average area under the receiver operating characteristic curve. Sens, the average sensitivity; Spec, the average specificity; J Stat, the average Youden’s J statistic; and AUC, the average area under the receiver operating characteristic curve. When comparing among the competing methods, we find the results are somewhat mixed across different settings. Specifically, SGL is the one we originally expected to outperform the rest of the competing methods as it also works under a hierarchical selection framework. Although SGL obtains higher sensitivities than the other competing methods in most of the scenarios, a lower specificity due to large number of false positives deteriorates the overall performance, especially in extreme situations. The two Bayesian methods piMASS and GEMMA achieve similar feature selection performance. Different from SGL, they lack power to detect true signals with very low sensitivities but high specificities. The performances of LASSO, SCAD, and FGWAS fluctuate in between, and none of them generally outperform the others. In terms of comparison among different settings, dimensions, and LD structures, as illustrated previously, methods generally work better to detect unified or clustered signal patterns but may encounter difficulty when both positive and negative effects exist. We further calculate the average marginal posterior probabilities for the risk SNPs in causal SNP-sets obtained from SGHS under both cases with . Figure 1 shows the boxplots representing these posterior probabilities for Setting 1 of Case 1 and Setting I of Case 2 under different dimensions. As shown in Figure 1, under Case 1, the possibilities to identify risk SNPs in each SNP-set are comparable. Under Case 2, where relatively sparse signals exist in the first five SNP-sets, we notice a higher probability in identifying signals within these SNP-sets. However, when it comes to SNP-sets 6–10 with diluted signals, the majority of genetic effect is lost during the selection. Finally, SGHS achieves an equally good, or even better, performance under high-dimensional scenarios compared with lower dimensions, and the impact of κ is not strong in our simulation settings.
Figure 1

Simulation results: the average marginal posterior probabilities of risk SNPs over each risk SNP-set under Setting 1 and Setting I with different dimensions.

Simulation results: the average marginal posterior probabilities of risk SNPs over each risk SNP-set under Setting 1 and Setting I with different dimensions.

The Alzheimer’s Disease neuroimaging initiative

There has been substantial interest in investigating neurodegenerative diseases such as Alzheimer’s disease (AD) based on neuroimaging and genetic markers. Data used in the preparation of this article were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu). The ADNI was launched in 2003 by the National Institute on Aging (NIA), the National Institute of Biomedical Imaging and Bioengineering (NIBIB), the Food and Drug Administration (FDA), private pharmaceutical companies, and nonprofit organizations, as a $60 million, 5-year public–private partnership. The primary goal of ADNI has been to test whether serial magnetic resonance imaging (MRI), positron emission tomography (PET), other biological markers, and clinical and neuropsychological assessment can be combined to measure the progression of mild cognitive impairment (MCI) and early AD. Determination of sensitive and specific markers of very early AD progression is intended to aid researchers and clinicians to develop new treatments and monitor their effectiveness, as well as lessen the time and cost of clinical trials. The Principal Investigator of this initiative is Michael W. Weiner, MD, VA Medical Center and University of California, San Francisco. ADNI is the result of efforts of many coinvestigators from a broad range of academic institutions and private corporations, and subjects have been recruited from over 50 sites across the US and Canada. The initial goal of ADNI was to recruit 800 subjects but ADNI has been followed by ADNI-GO and ADNI-2. To date, these three protocols have recruited over 1500 adults, aged 55–90 years, to participate in the research, consisting of cognitively normal older individuals, people with early or late MCI, and people with early AD. The follow-up duration of each group is specified in the protocols of ADNI-1, ADNI-2, and ADNI-GO. Subjects originally recruited by ADNI-1 and ADNI-GO had the option to be followed in ADNI-2. For up-to-date information, see www.adni-info.org. We conduct GWAS analysis on imaging phenotypes related to AD, and our goal is to identify genetic markers that are associated with imaging traits, and to further assess their predictive power. The advantage of using imaging phenotypes in GWAS is that imaging measurements tend to benefit the identification of pathogenic genes due to their close relationship with the biological etiology of multiple neurodegenerative and neuropsychiatric diseases, e.g., AD (Cannon and Keller 2006; Turner ; Peper ; Paus 2010; Scharinger ; Chiang ,b). For imaging traits, the raw MRI data were collected through 1.5 Tesla MRI scanners with protocols individualized for each scanner, including standard T1-weighted images obtained using volumetric three-dimensional (3D) sagittal MPRAGE or equivalent protocols with varying resolutions. The T1-weighted MRI images were preprocessed by standard steps including anterior commissure and posterior commissure correction, skull-stripping, cerebellum removing, intensity inhomogeneity correction, segmentation, and registration (Shen and Davatzikos 2004). Subsequently, 93 ROIs were labeled automatically by labeling the template and transferring the labels following the deformable registration of subject images (Wang ). After calculating the volume of each ROI for each subject, we consider nine of them as phenotypes: six subcortical regions, including left and right hippocampal volumes, left and right lateral ventricular volumes, and left and right amygdala volumes; and three global volumetric measures, including whole gray matter volume, whole white matter volume, and whole brain volume. A total of 818 subjects were genotyped using the Human 610-Quad BeadChip (Illumina, San Diego, CA). For data quality control, we focused on the 760 Caucasian subjects and removed ones identified as (i) sex check failure, (ii) >10% missing SNP, and (iii) outliers in the phenotypes/genotypes stratification, resulting in 745 subjects. The original SNPs data were generated from the Human Genome reference sequence build hg18, which were lifted over to hg19 in the current analysis. As a typical step in GWAS, we removed SNPs with (i) >5% missing values, (ii) minor allele frequency smaller than 5%, and (iii) Hardy-Weinberg equilibrium P-value . We also calculated the LD blocks to form the SNP-sets, and removed SNP-sets with single SNP. Eventually, 421,823 SNPs are left in our analysis, grouped into 16,084 SNP-sets with the number of SNPs varying from 2 to 100. We also include gender, age, and the first five principle component calculated by EIGENSOFT (Price ) into the analysis. We adopt the SGHS approach to investigate the joint association of SNPs with each of the nine MRI phenotypes in light of the autosomal LD blocks information. All the posterior simulation and hyper-parameters settings follow a similar line as the simulation studies. We first list all the selected SNP-sets associated with the nine imaging phenotypes along with the numbers of total SNPs and selected SNPs belonging to each set in Table 4. Based on the number of selected SNPs (columns 6 and 11), we observe the phenomenon that none of SNPs is selected within certain risk SNP-sets (the 0s in columns 6 and 11). This demonstrates the fact that some genetic information is diluted as it is widely studied in GWAS.
Table 4

ADNI data analysis results: list of selected SNP-sets associated with the phenotypes with the total number of SNPs and number of selected SNPs

ROIsChrBegin BPEnd BPTotal #aSelected #bChrBegin BPEnd BPTotal #Selected #
LA341021263412720813109121993508122159267378
42095663220827274440115005785455275456990
855302231554410253201221970019222429518621
88437355084826056590201288810512982672290
RA2810586058160745051085385167054119394502
6897416178998537972492062587521101230990
7104491613105158228575147322311073425315470
715016758315049108485111558109235349838976
81301362513253219997171887964919697976753
83222141232465554660
LH2381401263832830044061185380691191020359913
48967753790116432850108468049985193946900
48967753790116432850118590853786220724770
5106230898106426493250161262517112776281990
63871268838712688570
RH2333497170336237203203177565144177970694380
22122246892123857234804186063341186375488420
385591467862980879924857819845968366720
314675840514709360041488688695087362706730
LL51783033111784361903101152097415525951157924
61465592001469718485101290423439362931660
715157688154135745011154687680347310628740
RL1173854659175094025990114619884147293457800
51517392181524178679301610019899102920609942
632304085323950369935
GM2105454590105798292557681868051820610444721
6169471078169589925400
WM24653760446763587700898930457991098004819
6314341113151835499012106625131106950695510
847663704893353520203548715935925296330
WB61672877721675375945017947676774904969390

LA/RA, left/right amygdala volumes; LH/RH, left/right hippocampal volumes; LL/RL, left/right lateral ventricle volumes; GM, gray matter volume; WM, white matter volume; WB, whole brain volume.

Total number of SNPs in the SNP-set.

Number of selected SNPs in the SNP-set.

LA/RA, left/right amygdala volumes; LH/RH, left/right hippocampal volumes; LL/RL, left/right lateral ventricle volumes; GM, gray matter volume; WM, white matter volume; WB, whole brain volume. Total number of SNPs in the SNP-set. Number of selected SNPs in the SNP-set. We further consider the final SNP-level selection. The Manhattan plots in Figure 2 provide the SNP-wise inclusion probability with respective to each imaging phenotype. Under a 0.5 cutoff, for each phenotype, we map the selected SNPs to their associated genes, and summarize these risk genes along with the number of total/selected SNPs in Table 5. A further comparison between the number of selected SNPs and the total number of SNPs belonging to the risk gene (columns 3 and 4) demonstrates that our method is capable of identifying both sparse and diluted genetic information. Among the selected genes, a number of them have been reported previously in the literature. Such genes include ASAH2B (Avramopoulos ), SGMS1 (Hsiao ), GRIN2A (Leuba ), Tmem176b (Melchior ). In addition, several other genes have been shown to be related to the brain dysfunction or implicitly associated with Alzheimer’s disease. For instance, BRINP1 has been shown to highly express in various brain regions, and a lack of BRINP1 may lead to human psychiatric disorders (Kobayashi ). CCR6 has been implicated as an important biomarker associated with the inflammatory process of AD-like diseases (Subramanian ). RNASET2 deficiency interferes with brain development and myelination (Henneke ). Genes like CADM2, DLC1, and ABCC9 are related to Autism spectrum disorder (ASD) or Parkinson’s disease (Casey ; Jones ; Lin ), which may also serve as potential biomarkers for AD. Based on the selected genes, we also conduct a gene annotation analysis based on the enrichment for Kyoto Encyclopedia of Genes and Genomes (KEGG) (Kanehisa and Goto 2000). The associated pathways are calcium signaling pathway, which is a key component to regulate the neuronal excitability and processes related to the development neural diseases such as AD (Berridge 2013), and neuroactive ligand-receptor interaction, which is a well-known biomarker for cognitive ability (Antonell ; Kong ). As a comparison, we also perform GWAS based on single SNP analysis via PLINK (Purcell ) by performing quantitative trait association, and provide the Manhattan plots for value under each imaging phenotype in Figure 3. As a result, there are considerably less risk SNPs [associated with the human collagen alpha 1 (XIII) chain gene COL13A1] identified under both the well accepted threshold and the threshold suggested by Li compared with the result obtained by SGHS.
Figure 2

ADNI data analysis results: Manhattan plots for inclusion probabilities of all SNPs in all autosomes based on SGHS.

Table 5

ADNI data analysis results: list of selected genes associated with the phenotypes with the total number of SNPs and the number of selected SNPs

ROIsGeneTotal Number of SNPsNumber of Selected SNPsChromosome
Amygdala leftBRINP14569
ABCC9681612
CMAS6112
Amygdala rightALDH1A291415
DLC113978
GIMAP4527
GIMAP7437
KMT2E527
LHFPL313117
TMEM176B517
GABRR12516
Hippocampal formation leftCEP85L3346
SLC35F110316
PLN116
Hippocampal formation rightCADM2116123
Lateral ventricle leftA1CF14110
ASAH2B2110
SGMS1671710
AGMO6567
Lateral ventricle rightBTNL218126
C6orf10109146
GRIN2A1234016
Gray matter volumeCACNA2D1122217
MRPS9842
White matter volumeERICH5518
MATN244168
Whole matter volumeCCR6926
FGFR1OP826
RNASET2336
Figure 3

ADNI data analysis results: Manhattan plots of P-values of SNPs in all autosomes based on single SNP analysis.

ADNI data analysis results: Manhattan plots for inclusion probabilities of all SNPs in all autosomes based on SGHS. ADNI data analysis results: Manhattan plots of P-values of SNPs in all autosomes based on single SNP analysis. Finally, we assess the capacity of the selected genetic markers to predict imaging phenotypes using polygenic score (The International Schizophrenia Consortium 2009). Besides single SNP analysis and SGHS, we also apply SGL as another competing method. We use twofold cross validation by randomly splitting the dataset into equally sized ones, and perform the analyses on each as the training one. Under the corresponding testing sets, the Nagelkerke pseudo is calculated first for single SNP analysis based on the selected SNPs under different thresholds of P-values, and the grid of selected SNP numbers is further used as the thresholds for SGL and SGHS to obtain their scores. The final is averaged over two testing sets and we repeat the procedure five times to remove splitting bias. We present the results for all autosomes in Figure 4, which clearly shows a dramatic improvement of prediction by using SGHS compared with single SNP analysis and SGL in almost all the autosomes and thresholds. The predictive power for the selected risk profiles varies across different imaging phenotypes, and we also see a general pattern of an increase number of selected SNPs leading to a higher polygenic score.
Figure 4

ADNI data analysis results: Polygenic score under different thresholds in all autosomes based on single SNP analysis, SGL and SGHS.

ADNI data analysis results: Polygenic score under different thresholds in all autosomes based on single SNP analysis, SGL and SGHS.

Discussion

In this paper, we develop a unified Bayesian framework to realize hierarchical variable selection, while inducing grouping effect among predictors. Motivated by GWAS, our proposed method incorporates SNP-set information into the variable selection procedure, and facilitates selection at both SNP-set level and SNP level. Furthermore, by introducing a novel sampling scheme based on an auxiliary model for group-level selection, our approach is computationally efficient under high-dimensional feature space. We show in the simulation studies that the proposed method achieves considerably better performance than a number of competing methods under a wide range of settings. By applying the proposed method to the ADNI data set, we identify important genetic information that is highly associated with the volumes of ROIs in the brain. While our method is applied to an imaging-genetics study with a quantitative trait as phenotype, it is directly applicable to a dichotomous variable (e.g., case or control). As discussed in Albert and Chib (1993), one could use a probit regression model for the binary outcome, which leads to few modifications on the current posterior sampling scheme. In addition, we can also consider an incorporation of more biological information in the selection procedure. For instance, it is interesting to conduct Bayesian variable selection by incorporating the information on pathways and gene networks in microarray data (Li ; Stingo ) or functional connectivity for neuroimaging studies (Huang ; Goldsmith ). Similarly, we may introduce Ising or binary Markov random field (MRF) priors to the two levels of selection indicators in order to incorporate hierarchical biological information. After a realization of whole genome-wide association analysis, one extension of our work is to move forward to the whole-brain and whole GWAS. In this case, we need to use a multiple multivariate regression model to further capture the association among phenotypes (Zhu ). Besides the potential low detection power, the prohibitive computational cost will be the biggest issue of such analysis. A different direction is to extend the current model to the longitudinal data, which will increase the power to detect genetic association with neuroimaging phenotypes (Xu ). In this case, we need to modify our method to model the temporal association between responses and predictors while accounting for complex temporal correction structure.
  69 in total

1.  On the identification of disease mutations by the analysis of haplotype similarity and goodness of fit.

Authors:  Jung-Ying Tzeng; B Devlin; Larry Wasserman; Kathryn Roeder
Journal:  Am J Hum Genet       Date:  2003-02-27       Impact factor: 11.025

2.  Evaluating the effective numbers of independent tests and significant p-value thresholds in commercial genotyping arrays and public imputation reference datasets.

Authors:  Miao-Xin Li; Juilian M Y Yeung; Stacey S Cherny; Pak C Sham
Journal:  Hum Genet       Date:  2011-12-06       Impact factor: 4.132

3.  A variable selection method for genome-wide association studies.

Authors:  Qianchuan He; Dan-Yu Lin
Journal:  Bioinformatics       Date:  2010-10-29       Impact factor: 6.937

4.  Comparison of association mapping methods in a complex pedigreed population.

Authors:  Goutam Sahana; Bernt Guldbrandtsen; Luc Janss; Mogens S Lund
Journal:  Genet Epidemiol       Date:  2010-07       Impact factor: 2.135

5.  Robust deformable-surface-based skull-stripping for large-scale studies.

Authors:  Yaping Wang; Jingxin Nie; Pew-Thian Yap; Feng Shi; Lei Guo; Dinggang Shen
Journal:  Med Image Comput Comput Assist Interv       Date:  2011

6.  A principal components regression approach to multilocus genetic association studies.

Authors:  Kai Wang; Diana Abbott
Journal:  Genet Epidemiol       Date:  2008-02       Impact factor: 2.135

7.  Multiple SNP Set Analysis for Genome-Wide Association Studies Through Bayesian Latent Variable Selection.

Authors:  Zhao-Hua Lu; Hongtu Zhu; Rebecca C Knickmeyer; Patrick F Sullivan; Stephanie N Williams; Fei Zou
Journal:  Genet Epidemiol       Date:  2015-10-30       Impact factor: 2.135

8.  BDNF gene effects on brain circuitry replicated in 455 twins.

Authors:  Ming-Chang Chiang; Marina Barysheva; Arthur W Toga; Sarah E Medland; Narelle K Hansell; Michael R James; Katie L McMahon; Greig I de Zubicaray; Nicholas G Martin; Margaret J Wright; Paul M Thompson
Journal:  Neuroimage       Date:  2010-12-30       Impact factor: 6.556

9.  Dual induction of TREM2 and tolerance-related transcript, Tmem176b, in amyloid transgenic mice: implications for vaccine-based therapies for Alzheimer's disease.

Authors:  Benoit Melchior; Angie E Garcia; Bor-Kai Hsiung; Katherine M Lo; Jonathan M Doose; J Cameron Thrash; Anna K Stalder; Matthias Staufenbiel; Harald Neumann; Monica J Carson
Journal:  ASN Neuro       Date:  2010-07-12       Impact factor: 4.146

10.  Bayesian joint selection of genes and pathways: applications in multiple myeloma genomics.

Authors:  Lin Zhang; Jeffrey S Morris; Jiexin Zhang; Robert Z Orlowski; Veerabhadran Baladandayuthapani
Journal:  Cancer Inform       Date:  2014-12-07
View more
  2 in total

Review 1.  Genome-Wide Association Study Statistical Models: A Review.

Authors:  Mohsen Yoosefzadeh-Najafabadi; Milad Eskandari; François Belzile; Davoud Torkamaneh
Journal:  Methods Mol Biol       Date:  2022

2.  Bayesian interaction selection model for multimodal neuroimaging data analysis.

Authors:  Yize Zhao; Ben Wu; Jian Kang
Journal:  Biometrics       Date:  2022-02-27       Impact factor: 1.701

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.