Literature DB >> 35751586

Variational Bayes for high-dimensional proportional hazards models with applications within gene expression.

Michael Komodromos1, Eric O Aboagye2, Marina Evangelou1, Sarah Filippi1, Kolyan Ray1.   

Abstract

MOTIVATION: Few Bayesian methods for analyzing high-dimensional sparse survival data provide scalable variable selection, effect estimation and uncertainty quantification. Such methods often either sacrifice uncertainty quantification by computing maximum a posteriori estimates, or quantify the uncertainty at high (unscalable) computational expense.
RESULTS: We bridge this gap and develop an interpretable and scalable Bayesian proportional hazards model for prediction and variable selection, referred to as SVB. Our method, based on a mean-field variational approximation, overcomes the high computational cost of MCMC whilst retaining useful features, providing a posterior distribution for the parameters and offering a natural mechanism for variable selection via posterior inclusion probabilities. The performance of our proposed method is assessed via extensive simulations and compared against other state-of-the-art Bayesian variable selection methods, demonstrating comparable or better performance. Finally, we demonstrate how the proposed method can be used for variable selection on two transcriptomic datasets with censored survival outcomes, and how the uncertainty quantification offered by our method can be used to provide an interpretable assessment of patient risk.
AVAILABILITY AND IMPLEMENTATION: our method has been implemented as a freely available R package survival.svb (https://github.com/mkomod/survival.svb). SUPPLEMENTARY INFORMATION: Supplementary materials are available at Bioinformatics online.
© The Author(s) 2022. Published by Oxford University Press.

Entities:  

Year:  2022        PMID: 35751586      PMCID: PMC9364383          DOI: 10.1093/bioinformatics/btac416

Source DB:  PubMed          Journal:  Bioinformatics        ISSN: 1367-4803            Impact factor:   6.931


1 Introduction

The development of high-throughput sequencing technologies has led to the production of large-scale molecular profiling data, allowing us to gain insights into underlying biological processes (Widłak, 2013). One such technology is microarray sequencing, in which mRNA counts are used to describe gene expression. Such data, known as transcriptomics, are widely used in the biomedical domain and when analyzed alongside survival times have provided extraordinary opportunities for biomarker characterization and prognostic modeling (Bøvelstad ; Lightbody ; Lloyd ; Lu ). However, profiling data are often high-dimensional, which introduces several statistical challenges including: (i) variable selection, (ii) effect estimation of the features, (iii) uncertainty quantification and (iv) scalable computation. The task of variable selection is particularly important, as few genes typically have an effect on the outcome. Motivated by clinical applicability, we propose a state-of-the-art scalable (variational) Bayesian variable selection method for the proportional hazards model (PHM). In recent years, several methods have been proposed to analyze sparse high-dimensional data, with one of the most popular being the LASSO (Tibshirani, 1996). As biomedical studies are often concerned with clinical phenotypes, such as time to disease recurrence or overall survival time, these methods have been adapted to support survival analysis (Antoniadis ; Witten and Tibshirani, 2010). For instance, the LASSO, ridge and elastic-net penalties have all been extended to the PHM (Gui and Li, 2005; Simon ; Tibshirani, 1997; Zou and Hastie, 2005). More recently, Bayesian shrinkage and variable selection methods have grown in popularity (Bai ; Bhadra ; Carvalho ; Lewin ; Li and Zhang, 2010; O’Hara and Sillanpää, 2009; Park and Casella, 2008), with several methods being extended to survival data (Maity ; Nikooienejad ; Tang ). Bayesian approaches to variable selection are popular, not least since the relevance of a covariate can be assessed simply by computing the posterior probability that it is included in a model. This recasts variable selection as a model selection problem (George and McCulloch, 1993; Mitchell and Beauchamp, 1988), with every possible model assigned an individual posterior probability. One of the most popular such model selection priors is the spike-and-slab prior, see (Banerjee ) for a recent survey. However, exact posterior computation involves summing over models, where p is the number of covariates, which is intractable for even moderate p. Markov chain Monte Carlo (MCMC) methods avoid this problem but are known to have difficulty efficiently exploring the model space for high-dimensional covariates (Ormerod ), especially for the problem sizes found in many modern omics studies which motivate our work. This high computational cost has led to several methods either making continuous relaxations, giving rise to continuous shrinkage priors (Banerjee ; O’Hara and Sillanpää, 2009), or computing only maximum a posteriori (MAP) estimates, thereby not offering the full Bayesian machinery. Since we wish to preserve certain interpretable features arising from the original discrete model selection approach, such as inclusion probabilities of particular covariates for variable selection, we instead turn to variational inference. Variational inference (VI) is a popular scalable approximation technique, which has proven to be an effective tool for approximate Bayesian inference in many settings. VI involves minimizing the Kullback–Leibler divergence between a family of tractable distributions, called the variational family, and the posterior distribution; thereby recasting conditional inference as an optimization problem. The resulting minimizer is then used for downstream Bayesian inference. Though the approximation does not provide exact Bayesian inference, computationally convenient variational families can dramatically increase scalability. A common choice being mean-field families, under which the model parameters are independent. For a detailed review of VI, we direct the reader to Blei and Zhang . We propose a scalable and interpretable Bayesian PHM using a sparsity-inducing spike-and-slab prior with Laplace slab and Dirac spike, referred to as sparse variational Bayes (SVB). Since the posterior is computationally intractable, we use a mean-field variational approximation based on a factorizable family of spike-and-slab distributions, thereby preserving certain desirable discrete model selection aspects while providing scalable approximate Bayesian inference. We derive a coordinate-ascent algorithm for our implementation and investigate its performance in extensive simulations, comparing it against the posterior obtained via MCMC and demonstrating that the variational Bayes posterior can be used as a viable alternative, whilst being orders of magnitude faster to compute. We further compare with other state-of-the-art Bayesian variable selection methods, demonstrating comparable or better performance in many settings. Finally, we analyze two transcriptomic datasets involving ovarian and breast cancer data with censored survival outcomes, yielding biologically interpretable results. Various versions of this sparse variational family have been employed in linear and logistic regression models (Carbonetto and Stephens, 2012; Logsdon ; Ormerod ; Ray ; Ray and Szabó, 2021; Titsias and Lázaro-Gredilla, 2011) with some of these works specifically motivated by high-dimensional genomic applications. While most of these works use Gaussian distributions for the slab component, we instead follow Ray and Szabó (2021) in using a Laplace prior slab since Gaussian prior slabs are known to cause excessive shrinkage leading to potentially poor performance, even when exact posterior computation is possible (Castillo and van der Vaart, 2012). Our work can thus be viewed as extending ideas from the sparse VI literature to the setting of survival analysis under censoring. More generally, (not necessarily sparse) VI has proven to be an effective tool for approximate Bayesian inference and has seen wide use in several settings, including linear and logistic regression (Jaakkola and Jordan, 1996; Knowles and Minka, 2011), group factor analysis (Klami ), topic modeling (Blei and Lafferty, 2007), clustering (Teschendorff ) and Gaussian processes (Opper and Arachambeau, 2009) amongst others, with many of these methods employed in genomic and transcriptomic studies (Logsdon ; Papastamoulis ; Svensson ; Zhang and Flaherty, 2017).

2 Materials and methods

Notation: Let denote the observed data, where is an observed (possibly right censored) survival time, is a censoring indicator with if the observation is right censored and if the observation is uncensored, and is a vector of explanatory variables.

2.1 Survival analysis and the proportional hazards model

Let denote a random variable for an event time with density f(t) and cumulative distribution function F(t). Then the survival function, the probability a subject survives past time t, is given by where is the cumulative hazard rate and is the hazard rate, the instantaneous rate of failure at time t. Importantly, expressing S(t) in terms of the hazard function h provides a natural mechanism for analyzing survival times by assuming a form for h (Clark ; Ibrahim ). One such form, used to quantify the effect of features collected alongside survival times, is the proportional hazards model, wherein, where is a baseline hazard rate and are the model coefficients corresponding to the potential covariates of interest. Typically, estimating β is done by maximizing the partial likelihood, where (Cox, 1972, 1975). Under the partial likelihood, the baseline hazard rate is treated as a nuisance parameter and not specified, meaning the survival function is not directly accessible without further assumptions on the hazard rate. This approach is commonly used when the main interest is on quantifying the effect of covariates on the survival time to understand the underlying mechanisms, rather than purely for predictive purposes. Since our focus is on variable selection and analyzing effect sizes, we use the partial likelihood to compute the posterior. The use of the partial likelihood (3) is common in Bayesian survival analysis and can be understood via multiple Bayesian and frequentist justifications (Ibrahim ). For the frequentist, the partial likelihood is the empirical likelihood with the maximum likelihood estimator (MLE) for the cumulative baseline hazard function H0 plugged in, i.e. the profile likelihood (Murphy and Van Der Vaart, 2000). Using it in a Bayesian way thus means we are fitting a prior to our parameter of interest β and an MLE on the nuisance parameter H0. For the Bayesian, assigning a Gamma process prior to H0, marginalizing the posterior over H0 and taking the limit as the prior on H0 becomes non-informative, gives a marginal posterior for β exactly based on the partial likelihood (3) (Kalbfleisch, 1978). Thus using (3) can be viewed as using a diffuse Gamma process prior on the nuisance parameter H0.

2.2 Prior and variational family

We consider a spike-and-slab prior (George and McCulloch, 1993; Mitchell and Beauchamp, 1988) for the model coefficients β. Our choice of prior is conceptually natural for variable selection problems as it leads to interpretable inference regarding the inclusion probabilities of individual features. However, unlike the original formulation which uses Gaussian slabs, we use Laplace slabs, since Gaussian slabs are known to overly shrink the true large signals (Castillo and van der Vaart, 2012; Ray and Szabó, 2021). Formally, the prior distribution, , has hierarchical representation, where is a Dirac mass at zero, Laplace(λ) has density function on and . Placing a hyperprior on allows mixing over the sparsity level and allows adaptation to the unknown sparsity. The posterior density is proportional to the partial likelihood L in (3) times the joint prior density, formally, where and . Since the posterior (5) is computationally intractable, we use a variational approximation. For the variational family, we choose a mean-field family given by the product of independent spike-and-slab distributions with normal slab and Dirac spike for each coefficient: where . The notation ⊗ means a product measure implying coordinate independence, so that means Our choice of thereby provides scalability and maintains the property of variable selection via the Dirac mass, since the quantities are the inclusion probabilities. The variational posterior is then given by finding an element minimizing the KL divergence between Q and the posterior distribution , which is then used for inference. Note this approximation has O(p) parameters compared to the full posterior dimension . As with all mean-field approximation, dependent information between the components of β are lost, such as whether two coefficients β and β are likely to be selected simultaneously or not.

2.3 Coordinate-ascent algorithm

A convenient method for computing the mean-field variational posterior is coordinate-ascent variational inference (CAVI) (Blei ). In CAVI, the parameters for are sequentially updated by finding the values that minimize the KL divergence between the variational family and the posterior, whilst all other parameters are kept fixed, iterating until convergence. This reduces the overall optimization problem to a sequence of one-dimensional optimization problems. Minimizing the objective (7) is intractable for the Bayesian PHM due to the form of likelihood (3) and so we instead minimize an upper bound for the KL divergence. Such surrogate type functionals are well-used in variational inference, for example in logistic regression (Depraetere and Vandebroek, 2017; Jaakkola and Jordan, 1996; Knowles and Minka, 2011) and can lead to an increase in accuracy. The component-wise variational updates for μ and σ are given by the minimizers of and where and denotes the CDF of the standard normal distribution. The minimizers of these expressions do not have closed-form solutions, and therefore optimization routines are needed to find them, for instance via Brent’s method (Brent, 1973). Finally, the component-wise variational update for γ is given by solving, A full derivation of these expressions is provided in Supplementary Section SA. Algorithm 1 summarizes the CAVI algorithm. We denote the RHS of (10) by , and assess convergence by computing the change in and γ after each iteration, stopping when the total absolute change is below a specified threshold (e.g. ). While the evidence lower bound (ELBO) is often used to assess convergence, the ELBO is not analytically tractable in the present setting, which instead requires computationally expensive Monte Carlo integration to evaluate it. For this reason, we instead choose to assess convergence using the absolute change in and γ. Due to the non-convex objective in (7), CAVI generally only guarantees convergence to a local optimum, and therefore can be sensitive to initialization (Blei ). We found this to be the case for our method, particularly for μ and γ, therefore providing good starting values is generally important (see Supplementary Section SD for more details). In turn, we initialized μ using the LASSO with a small regularization hyperparameter, since μ corresponds to the unshrunk means if the variables are included in the model, and γ as , since this corresponds to an initial inclusion probability of 0.5 for each feature. We found the proposed method is less sensitive to initial value of σ, for example initializing σ as is sufficient. 1: require 2: Initialize 3: while not converged 4:    for 5:      // (8) 6:      // (9) 7:      // (10) 8: return.

2.4 Parameter tuning

The proposed method involves three prior parameters and b0 defined in (4), where λ controls the shrinkage imposed on , with large values imposing more shrinkage, and a0 and b0 control the shape of the Beta distribution, whose expectation reflects the a priori proportion of non-zero coefficients. Generally, our method is not particularly sensitive to the prior parameters (see Supplementary Section SD for a numerical investigation) and in practice using sensible a priori choices is appropriate for most settings. For example, if it is believed there are a small number of non-zero coefficients with moderate effect sizes, taking a0 as a small constant (such as ), and λ between 0.5 and 2.0 is appropriate. If an a priori choice is unavailable, the prior parameters can be tuned using the data. To do so, we suggest performing a grid search over a predefined set of values, selecting the element that maximizes a given goodness of fit measure, several options of which are presented in Supplementary Section SB. Furthermore, when tuning a0 and b0, to limit computation we suggest fixing b0 and searching across a set of values for a0, thereby exploring different values of the a priori inclusion probability.

2.5 Implementation

A freely available implementation is available for the R programming language via the package survival.svb, with functions available for fitting and evaluating models.

3 Simulation study

We use simulations to validate the proposed method, referred to as SVB. Firstly, we compare the variational posterior to the posterior obtained via MCMC, assessing whether our approximation can be used as a viable alternative. Secondly, we compare against other state-of-the-art Bayesian variable selection methods for the PHM. R scripts to reproduce our results can be found at https://github.com/mkomod/svb.exp.

3.1 Simulation design

Data are simulated for observations, each having a survival time t, censoring indicator δ and p continuous predictors . The survival time is sampled independently from , which has density , where we have taken and where the coefficient vector contains s non-zero elements with values sampled iid. uniformly from and indices chosen uniformly at random. To introduce censoring, we sample , letting where is the censoring proportion, and set where if , leaving t unchanged otherwise. Finally, the predictors are generated from one of four different settings designed to examine the behavior under varying degrees of difficulty: Setting 1, an independent setting where . Setting 2, a fairly challenging setting where predictors are moderately correlated within groups and independent between groups, formally with for otherwise. The setting is similar to Tang . Setting 3, a challenging setting where with estimated from the design of the TCGA dataset analyzed in Section 4.1. The s causal variables are randomly selected to correspond to features with a variance of at least 1.0. Setting 4, a realistic setting where the first p predictors are taken from the TCGA dataset analyzed in Section 4.1 and the s causal features are selected as in Setting 3. To evaluate the methods, we examine the accuracy of the corresponding point estimates, quality of the variables selected, and (if applicable) the uncertainty quantification. The point estimates are assessed via the and the , where is either a MAP estimate for β or the posterior mean if a distribution is available. For the variables selected the: (i) true positive rate (TPR) (ii) false discovery rate (FDR) and (iii) area under the curve (AUC) of the receiver operator characteristic curve are computed. For the TPR and FDR a coefficient is considered to have been selected if the posterior inclusion probability is at least 0.5. Finally, regarding uncertainty quantification, we evaluate the marginal credible sets by computing the: (i) empirical coverage, i.e. the proportion of times the true coefficient is contained in the credible set, and (ii) set size, given by the Lebesgue measure of the set. Details regarding the construction of the credible sets are presented where appropriate. For all metrics, we report the median, 5% and 95% quantiles across 100 replicates unless otherwise stated.

3.2 Simulation results

3.2.1 Comparison to MCMC

To assess how well the variational posterior matches the target (computationally challenging) posterior from (5), we compare the performance of our approach against the approximate yet asymptotically exact posterior obtained via MCMC. To do so, data is generated as described in Section 3.1, taking and , where we have kept n and p small so we can run our MCMC sampler in a reasonable amount of time. The MCMC sampler (described in Supplementary Section SC.1) was run for iterations with a burn-in period of iterations. For both methods, we used prior parameters and . Results are presented in Table 1.
Table 1.

Comparison of variational to MCMC posterior taking and , presented is the median and quantiles

Setting c Method 2 -error 1 -errorTPRFDRAUCRuntime
Setting 1 0.25SVB0.368 (0.21, 0.70)1.000 (0.52, 1.86)1.000 (0.90, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)18.5 s (13.5 s , 25.6 s)
MCMC0.412 (0.20, 0.75)1.017 (0.48, 2.01)1.000 (0.90, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)1 h 24 m (1 h 7 m , 1 h 50 m)
0.4SVB0.428 (0.23, 0.89)1.138 (0.63, 2.45)1.000 (0.90, 1.00)0.000 (0.00, 0.00)1.000 (0.95, 1.00)21.9 s (14.5 s , 30.5 s)
MCMC0.506 (0.26, 0.98)1.300 (0.69, 2.74)1.000 (0.80, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)1 h 28 m (1 h 25 m , 1 h 30 m)
Setting 2 0.25SVB0.376 (0.20, 0.73)1.031 (0.58, 2.07)1.000 (0.90, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)18.9 s (14.4 s , 25.4 s)
MCMC0.405 (0.21, 0.81)1.059 (0.58, 2.18)1.000 (0.90, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)1 h 14 m (1 h 6 m , 1 h 17 m)
0.4SVB0.472 (0.23, 1.08)1.176 (0.61, 2.96)1.000 (0.90, 1.00)0.000 (0.00, 0.00)1.000 (0.95, 1.00)24.0 s (17.3 s , 33.1 s)
MCMC0.520 (0.25, 1.08)1.319 (0.62, 2.91)1.000 (0.90, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)1 h 38 m (1 h 25 m , 2 h 4 m)
Setting 3 0.25SVB0.392 (0.18, 1.40)1.079 (0.53, 3.28)1.000 (0.90, 1.00)0.000 (0.00, 0.09)1.000 (0.95, 1.00)29.2 s (16.9 s , 44.9 s)
MCMC0.418 (0.21, 1.01)1.092 (0.54, 2.58)1.000 (0.90, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)1 h 45 m (1 h 24 m , 1 h 49 m)
0.4SVB0.470 (0.24, 1.57)1.263 (0.63, 4.16)1.000 (0.80, 1.00)0.000 (0.00, 0.10)1.000 (0.95, 1.00)21.7 s (13.7 s , 33.2 s)
MCMC0.508 (0.23, 1.26)1.236 (0.61, 3.45)1.000 (0.80, 1.00)0.000 (0.00, 0.09)1.000 (1.00, 1.00)1 h 36 m (1 h 30 m , 1 h 45 m)
Setting 4 0.25SVB0.393 (0.18, 1.12)1.067 (0.50, 2.54)1.000 (0.90, 1.00)0.000 (0.00, 0.10)1.000 (0.95, 1.00)17.0 s (9.2 s , 24.9 s)
MCMC0.382 (0.17, 0.95)1.007 (0.44, 2.47)1.000 (0.90, 1.00)0.000 (0.00, 0.10)1.000 (1.00, 1.00)1 h 5 m (1 h 3 m , 1 h 8 m)
0.4SVB0.425 (0.18, 1.38)1.171 (0.50, 2.85)1.000 (0.90, 1.00)0.000 (0.00, 0.10)1.000 (0.95, 1.00)25.8 s (14.8 s , 39.9 s)
MCMC0.486 (0.21, 1.13)1.158 (0.53, 3.17)1.000 (0.80, 1.00)0.000 (0.00, 0.00)1.000 (0.95, 1.00)1 h 38 m (1 h 14 m , 1 h 46 m)

Note: Simulations were ran on Intel® Xeon® E5-2680 v4 2.40 GHz CPUs.

Comparison of variational to MCMC posterior taking and , presented is the median and quantiles Note: Simulations were ran on Intel® Xeon® E5-2680 v4 2.40 GHz CPUs. Regarding the point estimates, for both the MCMC and the variational posteriors we took as the posterior mean, which for the latter is given by . Promisingly, both methods produce similar results, with near identical performance in all settings (Table 1). In particular, the similarity of the -error and -error suggests the posterior means are near identical. In terms of variable selection, both methods performed similarly. In particular, the TPR is comparable across the different settings, suggesting both methods are selecting a similar set of truly associated features. However, the upper quantile for the FDR is slightly larger for the variational posterior, meaning the MCMC posterior selects fewer spurious variables. Finally, we examine the uncertainty quantification of each method via 95% marginal credible sets , which are given by: S = I if the posterior inclusion probability is greater than 0.95, if the posterior inclusion probability is <0.05, and otherwise, where I is the smallest interval from the continuous component of our posterior such that S contains 95% of the posterior mass. As expected, for the non-zero coefficients, the coverage of the MCMC posterior is slightly better than the coverage of the variational posterior (Table 2), meaning the credible sets of the variational posterior are sometimes not large enough to capture the true non-zero coefficients. This is further reflected by the smaller set sizes, highlighting the well-known fact that VI can underestimate the posterior variance (Blei ; Carbonetto and Stephens, 2012; Ray ; Zhang ). Promisingly, the coverage of the zero coefficients is equal to one for both methods, meaning the credible sets contain zero, and typically, as reflected by the set size, contain only zero.
Table 2.

Coverage and set size for variational and MCMC posterior

Set. c Meth.Cov. β00Set size β00Cov. β0=0Set size β0=0
10.25SVB0.770 (0.202)0.320 (0.013)1.000 (0.000)0.000 (0.000)
MCMC0.928 (0.138)0.506 (0.039)1.000 (0.000)0.000 (0.000)
0.4SVB0.774 (0.208)0.355 (0.021)1.000 (0.000)0.000 (0.000)
MCMC0.914 (0.127)0.570 (0.054)1.000 (0.000)0.000 (0.000)
20.25SVB0.703 (0.227)0.306 (0.028)1.000 (0.001)0.000 (0.000)
MCMC0.904 (0.161)0.522 (0.053)1.000 (0.000)0.000 (0.000)
0.4SVB0.683 (0.262)0.333 (0.039)1.000 (0.001)0.000 (0.000)
MCMC0.845 (0.218)0.567 (0.101)1.000 (0.000)0.000 (0.000)
30.25SVB0.626 (0.288)0.251 (0.020)1.000 (0.000)0.000 (0.000)
MCMC0.903 (0.140)0.482 (0.047)1.000 (0.000)0.000 (0.000)
0.4SVB0.619 (0.278)0.276 (0.028)1.000 (0.000)0.000 (0.000)
MCMC0.873 (0.197)0.540 (0.078)1.000 (0.000)0.000 (0.000)
40.25SVB0.672 (0.224)0.252 (0.021)1.000 (0.000)0.000 (0.000)
MCMC0.921 (0.144)0.483 (0.047)1.000 (0.000)0.000 (0.000)
0.4SVB0.660 (0.249)0.277 (0.025)1.000 (0.001)0.000 (0.000)
MCMC0.906 (0.156)0.547 (0.059)1.000 (0.000)0.000 (0.000)

Note: Presented are means and std. dev.

Coverage and set size for variational and MCMC posterior Note: Presented are means and std. dev. Overall, the variational posterior displays similar performance to the MCMC posterior in key aspects for this setting with p = 1000 and can be computed orders of magnitude faster (Table 1). Our results highlight that the variational posterior is particularly good at capturing the key features (posterior means and inclusion probabilities) and provides reasonable uncertainty quantification for individual features.

3.2.2 Comparison to other methods

We perform a large-scale simulation study to empirically compare the performance of our method to two Bayesian variable selection methods. Within our study, data is generated as described in Section 3.1, taking and for all settings. Notably, under such a setting running MCMC would be computationally prohibitive, as highlighted in the previous section. We compare against BhGLM (Tang ), a spike-and-slab LASSO method that uses a mixture of Laplace distributions with one acting as the spike and the other the slab, and BVSNLP (Nikooienejad ), which uses a mixture prior composed of a point mass at zero and an inverse moment prior. Notably, both BhGLM and BVSNLP use Cox’s partial likelihood in the posterior and return a MAP estimate for β as well as posterior inclusion probabilities for each feature. Finally, for each method we use the default hyperparameters and let λ = 1, and for SVB. Generally, all methods produced excellent point estimates, with SVB obtaining the smallest median -error and -error in Settings 1 and 2, and BhGLM in Setting 4 (Table 3). Notably, SVB obtained the smallest lower (5%) quantile for the -error and -error in Settings 3 and 4, meaning the method can perform better than BhGLM but may be sensitive to the design matrix.
Table 3.

Comparison of Bayesian variable selection methods, taking and , presented is the median and quantiles

Setting c Method 2 -error 1 -errorTPRFDRAUC
Setting 1 0.25SVB0.378 (0.26, 0.89)1.747 (1.16, 4.17)1.000 (1.00, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)
BhGLM1.206 (0.79, 1.78)9.590 (7.22, 12.88)1.000 (1.00, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)
BVSNLP0.456 (0.33, 0.96)2.007 (1.41, 4.57)1.000 (1.00, 1.00)0.000 (0.00, 0.03)1.000 (1.00, 1.00)
0.4SVB0.449 (0.31, 0.99)2.056 (1.37, 4.87)1.000 (1.00, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)
BhGLM0.807 (0.53, 1.35)6.458 (4.52, 9.31)1.000 (1.00, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)
BVSNLP0.518 (0.35, 1.44)2.231 (1.52, 6.85)1.000 (0.96, 1.00)0.000 (0.00, 0.03)1.000 (0.99, 1.00)
Setting 2 0.25SVB0.405 (0.29, 0.80)1.823 (1.28, 3.78)1.000 (1.00, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)
BhGLM0.596 (0.45, 1.04)4.494 (3.51, 6.89)1.000 (1.00, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)
BVSNLP0.475 (0.33, 0.90)2.130 (1.47, 4.01)1.000 (1.00, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)
0.4SVB0.491 (0.33, 1.05)2.208 (1.45, 5.03)1.000 (0.97, 1.00)0.000 (0.00, 0.03)1.000 (1.00, 1.00)
BhGLM0.551 (0.44, 0.86)3.716 (2.98, 5.36)1.000 (0.97, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)
BVSNLP0.515 (0.37, 1.47)2.238 (1.54, 6.71)1.000 (1.00, 1.00)0.000 (0.00, 0.00)1.000 (1.00, 1.00)
Setting 3 0.25SVB1.040 (0.30, 3.17)3.881 (1.36, 15.37)0.967 (0.83, 1.00)0.000 (0.00, 0.14)0.983 (0.92, 1.00)
BhGLM0.590 (0.36, 1.57)3.279 (2.23, 6.73)1.000 (0.93, 1.00)0.000 (0.00, 0.03)1.000 (0.97, 1.00)
BVSNLP3.107 (1.74, 9.73)12.262 (6.88, 47.67)0.967 (0.53, 1.00)0.000 (0.00, 0.53)0.983 (0.78, 1.00)
0.4SVB1.379 (0.36, 3.47)5.728 (1.55, 17.47)0.933 (0.77, 1.00)0.033 (0.00, 0.13)0.967 (0.88, 1.00)
BhGLM0.796 (0.41, 2.18)4.035 (2.25, 10.92)0.967 (0.87, 1.00)0.000 (0.00, 0.07)1.000 (0.95, 1.00)
BVSNLP3.867 (1.98, 11.44)15.874 (7.99, 51.07)0.967 (0.20, 1.00)0.033 (0.00, 0.69)0.983 (0.65, 1.00)
Setting 4 0.25SVB0.603 (0.29, 2.02)2.298 (1.21, 8.84)1.000 (0.90, 1.00)0.000 (0.00, 0.08)1.000 (0.95, 1.00)
BhGLM0.503 (0.35, 1.36)3.141 (2.25, 5.59)1.000 (0.93, 1.00)0.000 (0.00, 0.03)1.000 (0.97, 1.00)
BVSNLP2.946 (1.96, 8.72)11.426 (6.98, 36.46)1.000 (0.90, 1.00)0.000 (0.00, 0.07)1.000 (0.95, 1.00)
0.4SVB1.092 (0.32, 2.83)3.878 (1.40, 14.06)0.967 (0.83, 1.00)0.000 (0.00, 0.08)0.983 (0.92, 1.00)
BhGLM0.674 (0.40, 1.64)3.610 (2.28, 7.72)1.000 (0.93, 1.00)0.000 (0.00, 0.04)1.000 (0.97, 1.00)
BVSNLP3.163 (2.14, 10.53)12.227 (8.14, 45.64)1.000 (0.73, 1.00)0.000 (0.00, 0.32)1.000 (0.87, 1.00)
Comparison of Bayesian variable selection methods, taking and , presented is the median and quantiles Regarding the variables selected, all methods performed exceptionally well achieving the ideal values for the TPR, FDR and AUC in Settings 1 and 2 (Table 3). Within Settings 3, BhGLM obtained the best TPR, FDR and AUC closely followed by SVB and BVSNLP. Within Setting 4, all three methods obtained the ideal values when the censoring was low (c = 0.25) and BhGLM performed best under moderate censoring (c = 0.40). Further, BhGLM best controlled the FDR in Settings 3 and 4, obtaining the lowest upper (95%) quantile, closely followed by SVB. Finally, we note, SVB is the only method that provides uncertainty quantification, a direct application of which is demonstrated in Section 4.2.

4 Application

4.1 TCGA ovarian cancer data

The first dataset we analyzed is a transcriptomic dataset where the outcome of interest is overall survival. The dataset was collected from patients with ovarian cancer and has a sample size of n = 580, of which 229 samples are right censored and 351 samples are uncensored, corresponding to a censoring rate of (TCGA, 2022). Within the dataset there are covariates, which we pre-processed by removing features with a coefficient of variation below the median value (Mar ), leaving 6021 covariates which we centered before fitting our method. When applying our method, we set and , reflecting our prior belief that few genes are associated with the response. As we had no prior belief for λ, we performed 10-fold cross-validation to select the value, exploring a grid of values . To evaluate model fit we compute the: (i) ELBO, (ii) expected log-likelihood under the variational posterior (ELL = ) and (iii) c-index, reporting the mean and standard deviation across the 10 folds for the training and validation sets in Supplementary Table S2. Notably, no single hyperparameter value stands out as being best, meaning the model is not particularly sensitive to the value of λ. To assess the model’s convergence diagnostics we examine the fit for λ = 1, and examine the change in: (i) ELBO, (ii) ELL and (iii) KL between the variational posterior and the prior, as we iterate our co-ordinate ascent algorithm (Fig. 1). Note the ELBO and ELL are computed for the training and validation sets, whereas need only be computed for the training set. Notably, across the different folds the ELBO is increasing as the co-ordinate ascent algorithm is iterated (Fig. 1A and B), suggesting that the model fit is improving. Interestingly, the training ELL is decreasing (Fig. 1C), whereas the inverse is true for the validation ELL (Fig. 1D), meaning, initially the model is overfitting to the training data, and as we iterate begins to fit better to the unseen validation set. Further, the is decreasing (Fig. 1E), therefore a greater degree of sparsity is enforced as we iterate, excluding more features and preserving the ones that best explain the variation in the response.
Fig. 1.

Ovarian cancer dataset model convergence diagnostics for λ = 1

Ovarian cancer dataset model convergence diagnostics for λ = 1 As we are using our model for variable selection, we examine the genes selected across the different values of λ and folds. Table 4 reports the names and selection proportion of genes, where the selection proportion is the number of times a particular gene has posterior inclusion probability greater than
Table 4.

Gene names and selection proportions for ovarian cancer dataset

PI3PPP3CACCR7SDF2L1D4S234EVSIG4DAPIL7R
0.70.3790.2930.2860.2290.1710.1360.136
TBPACSL3SLAMF7UBDIL2RGGALNT10FLJ20323RNF128
0.1210.1140.10.10.0640.0570.050.05
Gene names and selection proportions for ovarian cancer dataset Notably, is computed for each fit and is a threshold used to control the Bayesian FDR at significance level α, which we have set as (Newton ). Promisingly, the most frequently selected gene, PI3, has a known, albeit limited, role in ovarian cancer, a disease characterized by copy number aberration. Clauss reported the first link of PI3 gene product (elafin) to ovarian cancer (Clauss ). Elafin, a serine proteinase inhibitor involved in inflammation and wound healing, is overexpressed in ovarian cancer and overexpression is associated with poor overall survival and is due, in part, to genomic gains on chromosome 20q13.12, a locus frequently amplified in ovarian carcinomas. There is less known about the gene encoding the alpha isoform of the calcineurin catalytic subunit (PPP3CA) in ovarian cancer. A recent report indicates that higher expression of calcineurin predicts poor prognosis in ovarian cancer, particularly those of serous histology (Xin ). It is also plausible that other know functions of calcineurin/nuclear factor of activated T cells, in controlling adaptive T-cell function or innate immunity (Fric ), in this cancer that warrants further investigation. Finally, CCR7, the third most abundant gene was recently reported, in single cell RNA-seq analysis, to be emphasized in high-grade serous ovarian cancer (Izar ).

4.2 Breast cancer dataset

The second dataset we analyzed is a transcriptomic dataset collected from patients with breast cancer, where the outcome of interest is overall survival (Yau ). The dataset consists of n = 682 samples and p = 9168 features which we preprocessed as before, leaving p = 4, 584 features. Within the dataset 454 observations are right censored, corresponding to a censoring rate of 66.5%. As in the previous section, we set and , and tuned the prior parameter λ via 10-fold cross validation using the same set Λ. Supplementary Table S3 reports the ELBO, ELL and c-index averaged across the validation and training sets. We note that the model is not particularly sensitive to the value of λ. Furthermore, an assessment of the convergence diagnostics for , presented in Supplementary Figure S12, carries a similar interpretation as with the TCGA data. Table 5 reports the names and selection proportions of the genes within the dataset. The most frequently selected gene, Rho GTPase activating protein 28 (ARHGAP28) is a negative regulator of RhoA. There is paucity of data on this gene in cancer generally, however, a report by Planche identified the gene as downregulated in reactive stroma of prostate tumors. Further assessment of this gene in breast cancer is warranted. Notably, NEK2, GREM1 and ABCC5 have been examined in the biomedical literature and have been associated with cancer cell proliferation and metastasis. More specially, overexpression of NEK2 induces epithelial-to-mesenchymal transition, a process which leads to functional changes in cell invasion, overexpression of GREM1 has been associated with metastasis and poor prognosis, and ABCC5 has been associated with breast cancer skeletal metastasis (Mourskaia ; Park ; Rivera-Rivera ). As with the TCGA data, it is encouraging that genes with pre-existing biological interpretation have been selected by our model.
Table 5.

Gene names and selection proportions for the breast cancer dataset

ARHGAP28NEK2ABCC5GREM1DUSP4ITGA5CCL2IGFBP7
0.3860.250.20.20.1930.1930.1640.143
NFE2L3TRPC1PKMYT1DDX31EMILIN1SSPNABOHSPC072
0.1140.1140.10.0860.0860.0860.0790.079
Gene names and selection proportions for the breast cancer dataset Finally, we want to highlight that our method, in contrast to the methods compared in Section 3.2.2, quantifies the uncertainty of β. Crucially, the availability of uncertainty serves as a powerful inferential tool for computing (variational) posterior probabilities with respect to risk scores (. Such probabilities can be useful in comparing patients between one another, or assessing the risk of patients against chosen benchmarks (depending on the aims of the practitioner). To demonstrate, we opt to compute the posterior probability that one patient is at greater risk than another, formally, , where . To illustrate the application, we split patients into low- and high-risk groups based on the estimated prognostic index, , where is the posterior mean. Patients with prognostic index less than the median (computed for the training set) are considered low risk, whilst patients with prognostic index greater than or equal to the median are considered high risk. The Kaplan–Meier curves for these groups are shown in Figure 2A. Critically, Bayesian approaches that only compute the MAP of β are only able to provide a point estimates for . In turn, our method is able to provide uncertainty around this quantity and therefore with respect to the ranking of the patients. For instance, in Figure 2B, we present the posterior probabilities comparing the risks between patients. We observe that the highest risk patients in the low-risk group are comparable to the lowest risk patients in the high-risk group, and that the highest risk patients within the high-risk group are with high probability more at risk than the patients within the low-risk group.
Fig. 2.

(A) Kaplan–Meier curves for patients in low- and high-risk groups. (B) Comparison of patients in the low- and high-risk groups (ordered by )—within each cell the (variational) posterior probability patient in row i is at greater risk than patient in column j is computed. Samples are taken from the second validation fold and the fit with is used

(A) Kaplan–Meier curves for patients in low- and high-risk groups. (B) Comparison of patients in the low- and high-risk groups (ordered by )—within each cell the (variational) posterior probability patient in row i is at greater risk than patient in column j is computed. Samples are taken from the second validation fold and the fit with is used

5 Discussion

Variable selection and effect estimation for high-dimensional survival data has been an issue of great interest in recent years, particularly given the ever growing production of large scale omics data. However, the high-dimensionality and heterogeneity in the predictors, alongside the censoring in the response, makes the analysis a non-trivial task. While many recent methods have tackled these issues through a Bayesian approach, due to long compute times they often only produce point estimates rather than the full posterior and thereby fall short in providing the full Bayesian machinery. We have bridged this gap and developed a scalable and interpretable mean-field variational approximation for Bayesian PHMs with a spike-and-slab prior. We have demonstrated that the resulting variational posterior displays similar performance to the posterior obtained via MCMC whilst requiring a fraction of the compute time. Furthermore, we have demonstrated via an extensive simulation study that our proposed method performs comparably to state-of-the art Bayesian variable selection methods. Finally, we have demonstrated that our method can be used for variable selection on two real world transcriptomics datasets, giving rise to results with pre-existing biological interpretations, thereby validating the practical utility of our method. We have also shown that the risk of patients can be compared through (variational) posterior probabilities, highlighting that the availability of a posterior distribution can be a powerful inferential tool. For illustrative purposes, we examined the pairwise probabilities of patients grouped based on the prognostic index, however, patients could have alternatively been compared to other baselines, e.g. the feature vector corresponding to median prognostic index. Furthermore, although this is not an aspect we have considered, grouping based on: age, cancer status, stage etc. may yield insightful results for practitioners and bioinformaticians. A natural extension of our work would be to develop approximations with relaxed independence assumptions by using a more flexible variational family (Ning, 2021). Finally, we would like to highlight that improving the uncertainty quantification is an active area of research in the general VI community, see e.g. (Jerfel ). Click here for additional data file.
  31 in total

1.  Detecting differential gene expression with a semiparametric hierarchical mixture method.

Authors:  Michael A Newton; Amine Noueiry; Deepayan Sarkar; Paul Ahlquist
Journal:  Biostatistics       Date:  2004-04       Impact factor: 5.899

2.  Group Factor Analysis.

Authors:  Arto Klami; Seppo Virtanen; Eemeli Leppäaho; Samuel Kaski
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2014-12-18       Impact factor: 10.451

3.  A single-cell landscape of high-grade serous ovarian cancer.

Authors:  Benjamin Izar; Itay Tirosh; Elizabeth H Stover; Isaac Wakiro; Michael S Cuoco; Idan Alter; Christopher Rodman; Rachel Leeson; Mei-Ju Su; Parin Shah; Marcin Iwanicki; Sarah R Walker; Abhay Kanodia; Johannes C Melms; Shaolin Mei; Jia-Ren Lin; Caroline B M Porter; Michal Slyper; Julia Waldman; Livnat Jerby-Arnon; Orr Ashenberg; Titus J Brinker; Caitlin Mills; Meri Rogava; Sébastien Vigneau; Peter K Sorger; Levi A Garraway; Panagiotis A Konstantinopoulos; Joyce F Liu; Ursula Matulonis; Bruce E Johnson; Orit Rozenblatt-Rosen; Asaf Rotem; Aviv Regev
Journal:  Nat Med       Date:  2020-06-22       Impact factor: 53.440

4.  Identification of prognostic molecular features in the reactive stroma of human breast and prostate cancer.

Authors:  Anne Planche; Marina Bacac; Paolo Provero; Carlo Fusco; Mauro Delorenzi; Jean-Christophe Stehle; Ivan Stamenkovic
Journal:  PLoS One       Date:  2011-05-18       Impact factor: 3.240

Review 5.  Prediction of resistance to chemotherapy in ovarian cancer: a systematic review.

Authors:  Katherine L Lloyd; Ian A Cree; Richard S Savage
Journal:  BMC Cancer       Date:  2015-03-11       Impact factor: 4.430

6.  Variational inference for rare variant detection in deep, heterogeneous next-generation sequencing data.

Authors:  Fan Zhang; Patrick Flaherty
Journal:  BMC Bioinformatics       Date:  2017-01-19       Impact factor: 3.169

7.  Higher expression of calcineurin predicts poor prognosis in unique subtype of ovarian cancer.

Authors:  Bing Xin; Kai-Qiang Ji; Yi-Si Liu; Xiao-Dong Zhao
Journal:  J Ovarian Res       Date:  2019-08-09       Impact factor: 4.234

8.  The Nek2 centrosome-mitotic kinase contributes to the mesenchymal state, cell invasion, and migration of triple-negative breast cancer cells.

Authors:  Yainyrette Rivera-Rivera; Mihaela Marina; Shirley Jusino; Miyoung Lee; Jaleisha Vélez Velázquez; Camille Chardón-Colón; Geraldine Vargas; Jaya Padmanabhan; Srikumar P Chellappan; Harold I Saavedra
Journal:  Sci Rep       Date:  2021-04-27       Impact factor: 4.379

9.  ABCC5 supports osteoclast formation and promotes breast cancer metastasis to bone.

Authors:  Anna A Mourskaia; Eitan Amir; Zhifeng Dong; Kerstin Tiedemann; Sean Cory; Atilla Omeroglu; Nicholas Bertos; Véronique Ouellet; Mark Clemons; George L Scheffer; Morag Park; Michael Hallett; Svetlana V Komarova; Peter M Siegel
Journal:  Breast Cancer Res       Date:  2012-11-22       Impact factor: 6.466

10.  Discovery of a biomarker candidate for surgical stratification in high-grade serous ovarian cancer.

Authors:  Haonan Lu; Paula Cunnea; Katherine Nixon; Natasha Rinne; Eric O Aboagye; Christina Fotopoulou
Journal:  Br J Cancer       Date:  2021-01-21       Impact factor: 9.075

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.