Literature DB >> 36072060

Confirmation of the superior performance of the causal Graphical Analysis Using Genetics (cGAUGE) pipeline in comparison to various competing alternatives.

Richard Howey1, Heather J Cordell1.   

Abstract

Various methods exist that utilise information from genetic predictors to help identify potential causal relationships between measured biological or clinical traits. Here we conduct computer simulations to investigate the performance of a recently proposed causal Graphical Analysis Using Genetics (cGAUGE) pipeline, used as a precursor to Mendelian randomization analysis, in comparison to our previously proposed Bayesian Network approach for addressing this problem. We use the same simulation (and analysis) code as was used by the developers of cGAUGE, adding in a comparison with the Bayesian Network approach. Overall, we find the optimal method (in terms of giving high power and low false discovery rate) is the cGAUGE pipeline followed by subsequent analysis using the MR-PRESSO Mendelian randomization approach. Copyright:
© 2022 Howey R and Cordell HJ.

Entities:  

Keywords:  Bayesian networks; Causal inference; Mendelian randomization

Year:  2022        PMID: 36072060      PMCID: PMC9411974          DOI: 10.12688/wellcomeopenres.17991.1

Source DB:  PubMed          Journal:  Wellcome Open Res        ISSN: 2398-502X


Introduction

In a paper recently published by Amar , a pipeline entitled “causal Graphical Analysis Using Genetics” (cGAUGE) was proposed. This pipeline involves carrying out various pre-processing/filtering steps to reduce the number of variables to be taken forward for subsequent causal inference analysis using extensions of Mendelian randomization (MR) approaches, including inverse variance weighted (IVW) regression ( Burgess ), MR-Egger ( Bowden ) and MR-PRESSO ( Verbanck ). Amar demonstrated that use of their cGAUGE pipeline resulted in a lower false discovery rate (FDR) compared to carrying out the same MR analysis approaches with no pre-filtering of variables. Amar also considered the Bayesian network (BN) approach that we previously described ( Howey ). However, the BN methods were not optimally implemented by Amar , and only a single FDR value (rather than a detailed comparison of FDRs under different scenarios) was reported. We therefore here use the R simulation code of Amar and re-run some of their simulation examples to compare the MR methods (with/without cGAUGE pre-filtering) with our BN approach. We also take the opportunity to evaluate not just the FDR but also the power (at any given FDR) of the different methods considered, which was not reported by Amar . As previously discussed ( Howey ), there is a lack of direct comparability between MR and BN methods, and so for making discoveries with MR we chose to consider p-value threshold values of 0.1, 0.05 and 0.01, while for BN we used edge probability thresholds of 0.7, 0.8 and 0.9. The resulting powers and false discovery rates (FDR) are therefore not directly comparable, but they do give some indication of how the methods perform using thresholds that might be considered reasonably comparable choices in practice. In addition, by plotting receiver operating characteristic (ROC) curves, we can make direct comparisons between the methods in terms of the powers achieved at any given FDR.

Methods

To reproduce the simulated data and MR analyses of Amar , we downloaded the R code they provided. We used it to repeat the results presented in Figure 3a of Amar but we additionally analysed the data using a BN approach. We only considered the “UniqueIV” version of cGAUGE as this had been shown by Amar to perform the best. There were 15 continuous traits simulated, each having between 10 to 20 binary instrumental variables. The traits were set to have random causal relationships with around 1 or 2 other traits. Different levels of horizontal pleiotropy were set through assigning the simulation parameter p values of 0 , 0 .1,...0 .4, where zero indicates no horizontal pleiotropy. We used 100 simulation replicates and calculated the average false discovery rate (FDR) and the number of causal predictions between variables. We also calculated the average power of each method to identify the true causal relationships in each simulation. For MR, p-value threshold values of 0.1, 0.05 and 0.01, adjusted using the BY algorithm ( Benjamini & Hochberg, 1995; Benjamini & Yekutieli, 2001), were used. For BN, edge probability thresholds of 0.7, 0.8 and 0.9 were used.
Figure 3.

Close-up receiver operating characteristic (ROC) curves of Mendelian randomization and Bayesian network (BN) methods.

Data were simulated using the same methods and R code to reproduce the plots of Amar Figure 3a. The ROC curves show results from simulations for different levels of horizontal pleiotropy given by data simulation parameter p (see Amar ) set to either 0, 0.2 or 0.4. The p-values of the MR tests were adjusted using the BY algorithm in plots ( a) to ( c). IVW=inverse variance weighted.

In our earlier paper ( Howey ) we described two BN approaches. The first used only four variables (two traits and a weighted allele score variable for each trait) in any given analysis, and the second included all available trait variables simultaneously in the analysis, together with a weighted allele score variable for each trait variable. Here we reproduce Figure 3a of Amar , adding in our four-variable BN approach. (This is the approach that was labelled as BN(G,Z) in Figures 3–6 and as B1 in Figure 10 of Howey ). Weighted allele score variables were created using genetic instrumental variables (IVs) passing a p-value threshold of 0.05 when adjusted with Bonferonni correction from approximately 225 available IVs. We performed four-variable BN analyses for every pair of variables chosen from the 15 trait variables together with their accompanying weighted allele score variables. All variables were treated as continuous. We also separately carried out four-variable BN analyses using weighted allele score variables created from the subset of instrumental variables that were suggested by the “UniqueIV” cGAUGE method. All BN analyses were carried out using our freely available BayesNetty software package. The rationale for choosing the four-variable BN approach for each trait pair (instead of using all trait variables in one fitted network) was as follows: In the simulated data and related underlying network, Amar considered a variable to be causal on another variable if there was a directed path from one variable to the other with possible intermediate variables. Therefore, when an average BN is fitted to all the data, a fair comparison with the MR methods should account for the many potential paths when calculating the probability of a causal path from one variable to another. There are considerable practical challenges to correctly estimating the probability of a complete causal path from one variable to another from the average networks calculated by the BN approach when all the data is considered. The four-variable BN method avoids all such problems and has the advantage of a simple analysis approach, making it computationally extremely quick, particularly as no random restarts are required in the fitting process. For the BN analyses we calculated an average network for every pair of trait variables as previously described ( Howey ). To compute the average network, the data was bootstrapped with replacement 1000 times and the best fit network was fitted at each iteration. The weighted allele score variable for each trait variable was constrained to be fixed as a parent variable and no other edges were permitted to or from the weighted allele score variables. Thus, the only edge in question was between the two traits. The number of times that an edge appeared between two variables in each best fit network was recorded, together with its direction. This allows us to calculate the strength and direction values (between 0 and 1) for each pair of variables, where the strength is defined as the probability (proportion of times) that an edge appears between the two nodes and the direction is the proportion of times that the edge is in a given direction, given that it exists. The overall probability of a directed edge from one variable to another was given by the product of the strength and direction. We constructed ROC curves by calculating the mean true and false positive rates over the 100 simulation replicates for each test at different ( p-value or edge probability) thresholds. Three different levels of horizontal pleiotropy were considered, corresponding to p parameter values of 0, 0 .2 and 0 .4, respectively.

Results and discussion

Figure 1 shows the average FDR (leftmost panels), number of predictions (middle panels) and power (rightmost panels) for different p-value and edge probability thresholds ranging from the least stringent (top panels) to the most stringent (bottom panels). It can be seen by comparing the top left and top middle plots with the top panels of Amar Figure 3a, that we do indeed replicate their results (subject to some random stochasticity). As previously mentioned, the MR FDRs and powers cannot be compared directly with those of BN, but we apply edge probability thresholds that in practice could be considered reasonably similar and we thus consider it reasonable to compare these results.
Figure 1.

Mean number of discoveries, empirical false discovery rates (FDR) and power of Mendelian randomization and Bayesian network (BN) methods.

Data were simulated using the same methods and R code to reproduce plots from Amar , with the top left and top middle plots corresponding to Amar Figure 3a. The plots show the average results of the simulations for increasing levels of horizontal pleiotropy given by data simulation parameter p . Discoveries from each MR test were made at significance levels ( pval ) of 0.1, 0.05 and 0.01 in the top, middle and bottom rows respectively, after adjusting for FDR using the BY algorithm. Discoveries from each BN analysis were made at edge probability ( prob ) thresholds of 0.7, 0.8 and 0.9 in the top, middle and bottom rows respectively. The columns from left to right show the FDR, number of predictions and the power, respectively. IVW=inverse variance weighted.

Mean number of discoveries, empirical false discovery rates (FDR) and power of Mendelian randomization and Bayesian network (BN) methods.

Data were simulated using the same methods and R code to reproduce plots from Amar , with the top left and top middle plots corresponding to Amar Figure 3a. The plots show the average results of the simulations for increasing levels of horizontal pleiotropy given by data simulation parameter p . Discoveries from each MR test were made at significance levels ( pval ) of 0.1, 0.05 and 0.01 in the top, middle and bottom rows respectively, after adjusting for FDR using the BY algorithm. Discoveries from each BN analysis were made at edge probability ( prob ) thresholds of 0.7, 0.8 and 0.9 in the top, middle and bottom rows respectively. The columns from left to right show the FDR, number of predictions and the power, respectively. IVW=inverse variance weighted. When p = 0 there is no horizontal pleiotropy and the FDR for BN is slightly higher than that of other methods, but when p 0 the BN FDR is lower than MR-PRESSO (without use of cGAUGE) and around the same as IVW (without use of cGAUGE). Lower FDRs are achieved by the MR methods when “UniqueIV” cGAUGE filtering is applied, with MR-PRESSO combined with cGAUGE filtering achieving the overall lowest FDR. The number of predictions (middle panels) are included as these were presented in Amar , although we consider the power (rightmost panels) to be of more interest. The BN power is reasonably high but decreases when p increases, similar to what is seen for the MR methods with the exception of MR-PRESSO. We also tried using the IVs suggested by the cGAUGE method in our BN approach, even though the cGAUGE pipeline was originally designed for use with MR analyses. There was little difference in the FDR or power of BN between using cGAUGE or not. The FDR was around the same, a little higher when the edge probability threshold was 0.7 and a little lower when it was 0.9. The power was negligibly higher for all thresholds. The average FDRs obtained when using BNs with p = 0 and p = 0 .3 and using an edge probability threshold of 0.7 were 17% and 27%, respectively, similar to those reported by Amar . The FDR value for p = 0 was within the range suggested by Amar (between 16% and 21.4% among the top 20 predicted edges) but for p = 0 .3 the FDR was lower than that suggested by Amar ( > 31% among either the top 10 or top 20 predicted edges). Figure 2 shows the ROC curves obtained when p was set to 0, 0 .2 and 0 .4. The first obvious peculiar feature of the ROC curves in Figure 2 (a) to (c) is that none of the curves reach the top right hand corner (where the true and false positive rates would both equal 1). In BN analysis, on rare occasions the average network may never fit an edge between two traits if there is no strong association, resulting in a zero probability of an edge and thus a true or false positive rate of 1 is not possible for any probability threshold. The results of MR tests are adjusted using the BY algorithm ( Benjamini & Hochberg, 1995; Benjamini & Yekutieli, 2001) using the R package function p.adjust, to be consistent with the results of Amar . This adjustment results in many p-values that are truncated at 1 and thus do not pass any p-value threshold, resulting in no threshold where there is a true or false positive rate of 1. ROC curves in Figure 2 (c) to (d) show the MR methods without any adjustment, which are quite similar to those obtained when adjusting, but extend further to the top right hand corner.
Figure 2.

Receiver operating characteristic (ROC) curves of Mendelian randomization and Bayesian network (BN) methods.

Data were simulated using the same methods and R code to reproduce the plots of Amar Figure 3a. The ROC curves show results from simulations for different levels of horizontal pleiotropy given by data simulation parameter p (see Amar ) set to either 0, 0.2 or 0.4. The p-values of the MR tests were adjusted using the BY algorithm in plots ( a) to ( c) and not adjusted in plots ( d) to ( f). IVW=inverse variance weighted.

Receiver operating characteristic (ROC) curves of Mendelian randomization and Bayesian network (BN) methods.

Data were simulated using the same methods and R code to reproduce the plots of Amar Figure 3a. The ROC curves show results from simulations for different levels of horizontal pleiotropy given by data simulation parameter p (see Amar ) set to either 0, 0.2 or 0.4. The p-values of the MR tests were adjusted using the BY algorithm in plots ( a) to ( c) and not adjusted in plots ( d) to ( f). IVW=inverse variance weighted. The ROC curves in Figure 2 show the MR methods (particularly MR-PRESSO) that use the cGAUGE pipeline to be better in terms of providing the highest power at any given FDR, especially when there is horizontal pleiotropy, although they do have restricted true positive rates when the p-values are adjusted. Figure 3 shows a close-up view of the ROC curves from Figure 2 (focussing on small false positive rates) when the MR p-values are adjusted, in which it is easier to observe the overall optimality of the cGAUGE MR methods. When using the cGAUGE IV variables in the BN analyses, it can be seen from Figure 3 that there is an improvement in performance for low false positive rates, but, as shown in Figure 2 and Figure 3, the performance is worse for higher false positive rates.

Close-up receiver operating characteristic (ROC) curves of Mendelian randomization and Bayesian network (BN) methods.

Data were simulated using the same methods and R code to reproduce the plots of Amar Figure 3a. The ROC curves show results from simulations for different levels of horizontal pleiotropy given by data simulation parameter p (see Amar ) set to either 0, 0.2 or 0.4. The p-values of the MR tests were adjusted using the BY algorithm in plots ( a) to ( c). IVW=inverse variance weighted.

Conclusions

In conclusion, our results support the use of the cGAUGE pipeline and suggest that it should ideally be followed by subsequent MR-PRESSO analysis. This approach, at least in the scenarios considered here, generates the highest power at any given FDR, in comparison to competing approaches. However, as previously noted ( Howey ), BN analysis makes different assumptions from MR and offers a complementary alternative, as MR and BN methods may behave differently when different assumption violations occur. The BN approach has also been extended to incorporate medium-to-large amounts of missing data through a network-informed imputation approach ( Howey ), which can result in a substantial boost in power when applied to real large-scale data sets. This research was funded in whole, or in part, by the Wellcome Trust (Grant number 219424/Z/19/Z). For the purpose of open access, the author has applied a CC BY public copyright licence to any Author Accepted Manuscript version arising from this submission.

Data and software availability

Code to simulate data, along with the R implementation of cGAUGE, were downloaded from https://github.com/david-dd-amar/cGAUGE/. The BayesNetty software package is available from https://www.staff.ncl.ac.uk/richard.howey/bayesnetty/. In this paper, Howey and Cordell compare Mendelian Randomization (MR) -based methods, with and without the cGAUGE preprocessing pipeline, to their previously described Bayesian Network (BN) learning approach. The authors first replicate the simulation analysis done by Amar et al. 2021, and then extend the discussion by (1) improving the BN learning framework as compared to the implementation in Amar et al., and (2) presenting power and ROC calculations. The authors confirm the top performance of using cGAUGE with MR methods, but also provide further insights on when BNs perform well and how these methods are complementary. The paper is well written with detailed explanations of the analysis motivation and parameters. The problem of assessing causal inference methods is important, especially as some methods were repeatedly shown to have low performance (e.g., as shown by Amar et al. and others with respect to MR-Egger). This paper thus contributes by improving and extending the Amar et al. study, and by thoroughly discussing how to apply BN methods. I only have a few minor comments, as I list below: Typo: “using the R package function p.adjust” - > ‘“using the R function p.adjust” In the Conclusions section the authors nicely describe how MR and BN are complementary and also discuss their recent method for analysis of real data with missing values. These two points further strengthen the discussion about the utility of BN methods. These points can also be made in the abstract and the introduction to better orient the reader. For example, the authors can mention in the abstract that they also discuss how BNs offer a complementary alternative to MR as these approaches may behave differently when different assumption violations occur. Is the work clearly and accurately presented and does it cite the current literature? Yes If applicable, is the statistical analysis and its interpretation appropriate? Yes Are all the source data underlying the results available to ensure full reproducibility? Yes Is the study design appropriate and is the work technically sound? Yes Are the conclusions drawn adequately supported by the results? Yes Are sufficient details of methods and analysis provided to allow replication by others? Yes Reviewer Expertise: Computational Biology, Causality, Machine learning I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard. The manuscript by Howey & Cordell compares Graphical Analysis Using Genetics (cGAUGE) of Amar et al (2021) coupled with various Mendelian Randomisation (MR) techniques with an alternative approach of Bayesian network (BN) of Howey et al (2020). cGAUGE is a preprocessing framework to prepare data for the subsequent causal inference analysis, e.g. Mendelian randomization (MR) (which is used to establish causal links between various traits and factors). Authors intend a more thorough comparison of cGAUGE + various MR methods and BN, taking more care implementing BN than Amar et al did. Additionally, power of various methods is calculated. Howey & Cordell confirm that preceding MR analysis with cGAUGE yields lower FDR and that cGAUGE followed MR-PRESSO yields the optimal method. They, however, stress that direct comparison of MR and BN is non-trivial and that the two methods make different assumptions. The paper is well written and is thorough in describing implementation of the experiment, parameters and settings used and reasons behind their choices, code is available in a GitHub repository. Motivation and conclusions of the work are clearly stated and well discussed. It is a useful piece of work for people intending to use cGAUGE or BN for causal inference analysis. If I have to point out one thing it would be lack of comment/discussion of results from the point of view of the number of predictions and power --- whilst FDR results are extensively discussed, the other two measures of performance are not commented upon. Furthermore, it would be helpful to explain what one expects to observe in way of number of predictions --- wouldn't proportion of correctly identified causal links be a better measure? Is the work clearly and accurately presented and does it cite the current literature? Yes If applicable, is the statistical analysis and its interpretation appropriate? Yes Are all the source data underlying the results available to ensure full reproducibility? Yes Is the study design appropriate and is the work technically sound? Yes Are the conclusions drawn adequately supported by the results? Yes Are sufficient details of methods and analysis provided to allow replication by others? Yes Reviewer Expertise: machine learning, logic programming, statistics, genomics, quant gen I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.
  6 in total

1.  Mendelian randomization analysis with multiple genetic variants using summarized data.

Authors:  Stephen Burgess; Adam Butterworth; Simon G Thompson
Journal:  Genet Epidemiol       Date:  2013-09-20       Impact factor: 2.135

2.  Mendelian randomization with invalid instruments: effect estimation and bias detection through Egger regression.

Authors:  Jack Bowden; George Davey Smith; Stephen Burgess
Journal:  Int J Epidemiol       Date:  2015-06-06       Impact factor: 7.196

3.  Detection of widespread horizontal pleiotropy in causal relationships inferred from Mendelian randomization between complex traits and diseases.

Authors:  Marie Verbanck; Chia-Yen Chen; Benjamin Neale; Ron Do
Journal:  Nat Genet       Date:  2018-04-23       Impact factor: 38.330

4.  Graphical analysis for phenome-wide causal discovery in genotyped population-scale biobanks.

Authors:  David Amar; Nasa Sinnott-Armstrong; Euan A Ashley; Manuel A Rivas
Journal:  Nat Commun       Date:  2021-01-13       Impact factor: 14.919

5.  A Bayesian network approach incorporating imputation of missing data enables exploratory analysis of complex causal biological relationships.

Authors:  Richard Howey; Alexander D Clark; Najib Naamane; Louise N Reynard; Arthur G Pratt; Heather J Cordell
Journal:  PLoS Genet       Date:  2021-09-29       Impact factor: 5.917

6.  Bayesian network analysis incorporating genetic anchors complements conventional Mendelian randomization approaches for exploratory analysis of causal relationships in complex data.

Authors:  Richard Howey; So-Youn Shin; Caroline Relton; George Davey Smith; Heather J Cordell
Journal:  PLoS Genet       Date:  2020-03-02       Impact factor: 5.917

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.