Literature DB >> 28137713

HIPred: an integrative approach to predicting haploinsufficient genes.

Hashem A Shihab1, Mark F Rogers2, Colin Campbell2, Tom R Gaunt1.   

Abstract

MOTIVATION: A major cause of autosomal dominant disease is haploinsufficiency, whereby a single copy of a gene is not sufficient to maintain the normal function of the gene. A large proportion of existing methods for predicting haploinsufficiency incorporate biological networks, e.g. protein-protein interaction networks that have recently been shown to introduce study bias. As a result, these methods tend to perform best on well-studied genes, but underperform on less studied genes. The advent of large genome sequencing consortia, such as the 1000 genomes project, NHLBI Exome Sequencing Project and the Exome Aggregation Consortium creates an urgent need for unbiased haploinsufficiency prediction methods.
RESULTS: Here, we describe a machine learning approach, called HIPred, that integrates genomic and evolutionary information from ENSEMBL, with functional annotations from the Encyclopaedia of DNA Elements consortium and the NIH Roadmap Epigenomics Project to predict haploinsufficiency, without the study bias described earlier. We benchmark HIPred using several datasets and show that our unbiased method performs as well as, and in most cases, outperforms existing biased algorithms.
AVAILABILITY AND IMPLEMENTATION: HIPred scores for all gene identifiers are available at: https://github.com/HAShihab/HIPred . CONTACT: h.shihab@bristol.ac.uk or tom.gaunt@bristol.ac.uk. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.
© The Author 2017. Published by Oxford University Press.

Entities:  

Mesh:

Substances:

Year:  2017        PMID: 28137713      PMCID: PMC5581952          DOI: 10.1093/bioinformatics/btx028

Source DB:  PubMed          Journal:  Bioinformatics        ISSN: 1367-4803            Impact factor:   6.937


Introduction

Technological advances and the falling costs of next-generation sequencing technologies have accelerated the identification of genetic variation in the human genome (The 1000 Genomes Project Consortium, 2012). The most common form of genetic variation is single nucleotide variants (SNVs) and small insertions/deletions (INDELs). Identifying which of these are functional promises to improve our understanding of the molecular mechanisms of human disease and lead to novel treatments. As a result, there is a plethora of in silico algorithms capable of predicting the functional impact of SNVs and INDELs, e.g. (Choi ; Kircher ; Ritchie ; Shihab , 2015). On the other hand, loss-of-function (LoF) variants, i.e. truncating mutations and whole gene deletions, have traditionally been considered to be rare and highly deleterious. However, there is growing evidence to suggest that LoF variants are common amongst healthy individuals (MacArthur ; Ng ; Pelak ). Haploinsufficiency, whereby a single copy of a gene product is not sufficient to maintain the normal function of the gene, is just one possible biological mechanism implicating LoF variants to abnormal phenotypes (Veitia and Birchler, 2010). Prediction of haploinsufficiency is an important aspect to interpreting whole genome sequence data, in which each individual will have a number of nonsense and missense mutations. Therefore, accurate methods for identifying haploinsufficiency within the genome are of increasing importance. A large proportion of the existing algorithms for predicting haploinsufficiency utilize biological networks, such as protein–protein interaction networks. However, it has been shown that commonly used biological networks are heavily affected by study bias (Steinberg ); i.e. well studied genes are over-represented with respect to the number of networks they are part of and the number of links they form within these networks. As a result, these methods tend to perform best on well-studied genes but underperform on less studied genes. Steinberg constructed an unbiased genome-wide haploinsufficiency score (GHIS) by replacing these biological networks with co-expression networks. However, other potentially informative sources for functional annotation include the Encyclopaedia of DNA Elements (ENCODE) consortium (The ENCODE Project Consortium, 2012) and the NIH Roadmap Epigenomics Project (Roadmap Epigenomics Consortium ). Following our previous work (Shihab ), we describe a machine learning approach (called HIPred) that integrates genomic and evolutionary features with functional annotations from ENCODE and NIH Roadmap Epigenomics to predict haploinsufficiency. We observe improved performance when compared with five existing methods, but without the potential study bias described earlier. Pre-computed HIPred scores for all gene identifiers (GRCh37) are available at: https://github.com/HAShihab/HIPred.

2 Materials and methods

2.1 Datasets

We assembled two datasets for training: our positive dataset was constructed using 299 known haploinsufficient genes from (Dang ) and our negative dataset was constructed using 386 putative loss-of-function tolerant (LoFT) genes from (MacArthur ). After removing records with conflicting annotations and ambiguous mappings, we retained 298 haploinsufficient and 386 LoFT genes. Following a similar procedure described in Steinberg , we used the following benchmarks from (Petrovski ) to evaluate the performance of HIPred: 175 genes listed as haploinsufficient in OMIM (OMIM HI), 108 genes listed as haploinsufficient with known de novo mutations in OMIM (OMIM HI de novo), 91 genes for which a heterozygous gene knockout causes ‘lethality’ phenotypes in mouse (MGI Lethality) and 95 genes for which a heterozygous gene knockout causes seizures in mouse (MGI Seizures). Next, we collected a list of 59 genes disrupted by de novo LoF mutations in autism probands (ASD1) (Iossifov ) and a further 64 genes disrupted by de novo LoF mutations in other sets of autism probands (ASD2). (Neale ; O’Roak ; Sanders ). The composition of haploinsufficient genes across these benchmarks, and their overlap with our training data, is summarized in Supplementary Table S1. The actual genes used in these benchmarks are given as Supplementary Material. Finally, for each gene, we also obtained the number of associated publications in PubMed using the NCBI Entrez Search and Retrieval System and used this as a measure of how ‘well-studied’ these genes are.

2.2 Feature groups

Following our previous work (Shihab ), we annotated our datasets using a number of feature groups, which could be predictive of haploinsufficiency. A detailed description of these feature groups can be found in Supplementary Table S2, but a short description is as follows: Genomic and evolutionary: we used a number of genomic properties such as the length of the gene, number of transcripts and the average number of predicted protein domains across transcripts. A comprehensive set of conservation-based measures, such as dN/dS ratios between human and 65 different species (one-to-one orthologues), was also used. In addition, we also tested whether the number of observed rare variants (MAF < 0.01) from the Exome Aggregation Consortium (ExAC) (Lek ), the number of expected rare variants across the gene, and a z-score representing the deviation of observed from expected added any predictive value. Histone modifications: we used regions of enrichment based on histone ChIP-seq peak calls from ENCODE and NIH Roadmap Epigenomics. Open chromatin: we used regions of enrichment based on DNase-seq and Formaldehyde-Assisted Isolation of Regulatory Elements peak calls from ENCODE and NIH Roadmap Epigenomics. Transcription factor-binding sites: based on PeakSeq and SPP peak calls for 119 transcription factors across 77 cell lines from ENCODE. Gene expression: based on RNA-seq signal coverage using consolidated epigenomes from NIH Roadmap Epigenomics. Methylation: based on whole genome bisulphite sequencing from NIH Roadmap Epigenomics. Digital genomic footprinting sites: for transcription factor recognition sequences within DNase-hypersensitive sites using consolidated epigenomes from the NIH Roadmap Epigenomics Project. Networks: we used measures of centrality from cell-type specific interactome and tissue-specific co-expression networks. As described in the Supplementary Material, the majority of our feature groups comprise multiple annotations across a gene. For example, there could be multiple ChIP-seq values across a given region (one value for each position). In these instances, we used the median value across the region. We also tested other summary measures, specifically the mean and maximum value across a region; however, these summaries did not yield any significant improvements in the overall performance of our method (data not shown).

2.3 Data integration

The resulting product of our data preparation is several large matrices comprising data from the above feature groups, each of which can have different measurement scales. Therefore, we investigated three approaches for data integration (see Fig. 1). First, we evaluated data integration at the data level (i.e. concatenating datasets into a single matrix). This form of data integration is the simplest and most intuitive; however, combining feature groups in this way creates additional analytical challenges. For example, classifiers will need to handle a larger number of heterogeneous features. Therefore, we used a gradient boosted machine (Chen and Guestrin, 2016) as they can handle heterogeneous datasets, are robust to missing data and can estimate the relative importance of features. To illustrate the potential benefits of using a gradient boosted machine on this type of data, as opposed to alternative machine learning algorithms, we also evaluated the performance of a support vector machine (SVM) (Campbell and Ying, 2011) on the same task.
Fig. 1

Methods for integrating feature groups: (a) feature groups are combined at the data level and fed into a single classifier; (b) feature groups are encoded as base kernels and combined using MKL; and (c) feature groups are used to construct heterogeneous base classifiers which are then combined using a stacking approach

Methods for integrating feature groups: (a) feature groups are combined at the data level and fed into a single classifier; (b) feature groups are encoded as base kernels and combined using MKL; and (c) feature groups are used to construct heterogeneous base classifiers which are then combined using a stacking approach Next, we evaluated data integration based on multiple kernel learning (MKL). In MKL, feature groups are encoded into a corresponding base kernel (where if there are p feature groups), from which we can derive a composite kernel matrix . This composite kernel can then be used with a kernel-based classifier such as the SVM, which was the classifier used here. The are kernel weights where and . These weights can be adjusted according to the relative informative-ness of the different feature groups. We used an L1-norm to yield sparse solutions that implicitly excludes uninformative feature groups by assigning them zero weight. Finally, we evaluated data integration based on stacking. Here, each feature group was tested against a number of machine learning algorithms, e.g. naïve Bayes, SVMs and random forests, and the best performing algorithm was chosen as the base classifier for the group (where if there are p feature groups). These base classifiers were then ’stacked’ (i.e. combined) using a logistic regression: , where the β of each base classifier was deduced through the regression process. As with MKL, we used an L1-norm to implicitly exclude uninformative feature groups by assigning them zero coefficient. We present our results using several performance statistics, such as the overall accuracy, sensitivity and specificity. In addition, we provide receiver operating characteristic (ROC) curves and area under the curve (AUC) statistics. Individual algorithm parameters, e.g. the SVM cost parameter C, were optimized through a 10-fold cross-validation and grid search. To remove the potential bias caused by the random partitioning of the datasets during cross validation, we repeated our analysis 30 times and report the mean values and SDs above 0.01. In order to alleviate any performance artifacts arising from potential gene similarity within our training dataset, we performed a gene similarity analysis using NCBI’s BLASTCLUST algorithm using the following parameters: -p F, -L 0.6, -b F and -S 10. Finally, we performed a feature selection analysis to identify important features and improve model interpretation. All analyses in this study were performed using scikit-learn (Pedregosa ), SHOGUN (Sonnenburg ) and xgboost (Chen and Guestrin, 2016).

2.4 Sequential learning

Both MKL and stacking rely on the assumption that data are available in every feature group for every training example. However, in practice, data can be absent from some feature groups, e.g. not every position in the genome will have annotations for Open Chromatin. We could use all available feature groups, but this would result in fewer training examples. With all values present for all feature groups, our training dataset consisted 156 haploinsufficient genes and just 52 LoFT genes. On the other hand, we could use a smaller number of feature groups, but with many more training examples. Therefore, we opted for an iterative sequential learning approach to determine the best combination of base kernels (MKL) and classifiers (stacking) (Rogers ). Here, we rank the base kernels/classifiers based on their individual cross-validation performance. Then, starting with the best performing base kernel/classifier, we iteratively add feature groups and retest the combined model using the same cross-validation procedure. If performance improves, then the feature group is added to the final model and the process is repeated until no more feature groups can be added.

2.5 Comparison with existing methods

For each of the benchmarks described in Datasets, we compared HIPred with 5 alternative methods that could be used to predict haploinsufficiency: predicted haploinsufficiency probabilities, HIS and HIS Imputed (Huang ); predicted gene indispensability scores (IS) (Khurana ); Residual Variance Intolerance Scores (RVIS) (Petrovski ); Evolutionary Intolerance (EvoTol) (Rackham ); and predicted genome-wide haploinsufficiency probabilities (GHIS) (Steinberg ). In addition, we evaluated the correlation between each of the methods evaluated based on the absolute Spearmando rank correlation coefficient.

3 Results

3.1 Performance of the method

On our training dataset, the performance of existing methods ranged from 0.6929 to 0.8549, with the HIS (Imputed) probabilities achieving the highest AUC (see Table 1). However, due to potential overlaps between this dataset and the datasets used to train these algorithms, the performances reported here may be over-inflated and may not represent the true generalizability of existing methods (see section 3.2 below).
Table 1

Performance of haploinsufficiency predictors on our training data

MethodAccuracySensitivitySpecificityPrecisionNPVAUC
EvoTol0.63670.55770.79880.69050.69170.6929
GHIS0.70690.71780.63270.65780.69510.7450
RVIS0.81290.78950.75960.70590.83160.8329
HIS0.67070.66830.83830.83540.67310.8412
IS0.84780.84030.70170.67790.85470.8489
HIS (Imputed)0.61950.51550.92570.85810.68670.8549
HIPreda0.90320.88460.89190.85190.91670.8940

Note: NPV, negative predictive value; AUC, area under the curve.

The reported performance of HIPred is the average performance observed across our repeated cross-validation procedure.

Performance of haploinsufficiency predictors on our training data Note: NPV, negative predictive value; AUC, area under the curve. The reported performance of HIPred is the average performance observed across our repeated cross-validation procedure. Next, we evaluated the performance of a gradient boosted machine, i.e. data integration at the data level. In terms of AUC, the performance of our gradient boosted machine outperformed all existing methods with an average AUC of 0.8940. Comparing the performance of a gradient boosted machine and SVMs, we achieved a nominal AUC of 0.8133 using SVMs; thereby highlighting the potential pitfalls of integrating large heterogeneous datasets at the data level. In our experiments, the highest performing MKL model comprised seven feature groups and achieved an average AUC of 0.8747. Here, Genomic and Evolutionary was the highest performing individual feature group with an average AUC of 0.8179, followed by Open Chromatin and Histone Modifications from the NIH Roadmap Epigenomics Project (i.e. gappedPeak and narrowPeak) with an average AUC of 0.8103 and 0.8035, respectively. Histone Modifications from ENCODE yielded an average AUC of 0.7518. We observed a performance boost of 4.61% during the first stage of our sequential learning approach. However, we observed minor improvements at each subsequent iteration. Interestingly, MKL assigned the largest weight to Histone Modifications from ENCODE (0.6056), whose individual performance was ranked 4th overall, followed by Genomic and Evolutionary (0.2233) and Open Chromatin/Histone Modifications (narrowPeak) from NIH Roadmap Epigenomes Project (0.1701). A lower weight was assigned to gappedPeak and broadPeak (0.0008 and 0.0001, respectively), probably because of the similarities between these feature groups and the narrowPeak feature group (see Supplementary Tables S3 and S4). The best performing stacked model comprised four feature groups and had an average AUC of 0.8866. As with MKL, Genomic and Evolutionary was the best performing feature group with an average AUC of 0.8196, followed by Open Chromatin and Histone Modifications (narrowPeak) with an AUC of 0.8794. However, in contrast to MKL, we observed a small performance boost of 0.38% after the initial sequential learning iteration. In our experiments, the most informative feature group was Genomic and Evolutionary with a regression coefficient of 4.0810, followed by Gene Expression and Open Chromatin (regression coefficients 2.1805 and 2.1391, respectively). Despite the best performing model comprising four feature groups during cross-validation, the final logistic regression assigned zero coefficients to one of these feature groups (see Supplementary Tables S5 and S6). From our analysis, it would appear that all 3 data integration classifiers evaluated outperform existing methods, with a classifier based on integration at the data level performing best. However, the difference in performance between these three data integration techniques evaluated is marginal. Nonetheless, for maximum performance and improved model interpretation, the final version of HIPred is based on a gradient boosted machine. All subsequent analyses presented are based on this version. Next, we tested for potential gene similarity (at the nucleotide level) within our training data using the NCBI BLASTCLUST algorithm. Using a minimum sequence identity and sequence coverage of 60%, we did not find any gene clusters. Finally, we performed a feature selection analysis to identify the most informative features (see Fig. 2). The most important feature identified from this analysis was the ExAC (E) missense z-scores for the deviation of observed missense variants from expectation (gain = 0.34), followed by several cell-type specific interactomes (I), such as the Mesenchymal Precursor (gain = 0.09), and genomic/evolutionary features (G), such as the dN/dS ratios and percent identity with other Ensembl genomes. We also assessed the performance of our final model using a progressive number of features and found that a maximum tree depth of 2 yields optimal performance (see Supplementary Material).
Fig. 2

Informative features used for predicting haploinsufficient genes

Informative features used for predicting haploinsufficient genes

3.2 Performance using known and candidate disease genes

We evaluated HIPred on a set of known human disease-associated genes and mouse model equivalents. After removing genes that were present in our training data (see Supplementary Table S1), we were left with 81 OMIM HI and 49 OMIM HI de novo genes. Following the procedure described in Steinberg , we matched these genes to an equal number of random genes based on gene length, which were assumed to be neutral. In general, HIPred outperformed the RVIS, EvoTol and GHIS across both OMIM datasets (see Table 2). Furthermore, HIPred marginally outperformed the HIS (both imputed and non-imputed) and Indispensability probabilities. However, these OMIM datasets comprise some of the most studied genes which could artificially inflate the observed performances of biased predictors. ROC curves are made available as Supplementary Figure S1.
Table 2

Performance of methods used for predicting haploinsufficiency on known disease genes and mouse models

MethodAccuracySensitivitySpecificityPrecisionNPVAUC
OMIM HI
 EvoTol0.52320.52630.73580.74070.52000.6477
 GHIS0.80770.86300.36210.63000.67740.6845
 RVIS0.75930.83540.28070.61680.55170.6609
 HIS0.66040.70490.69230.78180.60000.7303
 IS0.78690.83540.51720.70210.69770.7451
 HIS (Imputed)0.49330.47220.83330.80950.51280.7156
 HIPred0.76060.78210.60260.66300.73440.7543
OMIM HI de novo
 EvoTol0.54550.54550.72730.72730.54550.6959
 GHIS0.83610.88890.29730.60610.68750.7135
 RVIS0.86670.91490.25000.61430.69230.6965
 HIS0.71880.75680.69230.77780.66670.7599
 IS0.82860.87230.48570.69490.73910.7350
 HIS (Imputed)0.54550.53490.83330.82140.55560.7357
 HIPred0.89190.91300.52170.65620.85710.7902
MGI lethality
 EvoTol0.49280.50000.71740.72920.48530.6258
 GHIS0.75760.82350.39580.65880.61290.6725
 RVIS0.66970.76000.36360.67060.47060.6523
 HIS0.56000.59260.77420.82050.52170.7210
 IS0.69490.75680.52000.70000.59090.7065
 HIS (Imputed)0.46760.44780.85370.83330.48610.7632
 HIPred0.78720.79730.70270.72840.77610.8143
MGI seizures
 EvoTol0.53410.52870.71640.70770.53930.6611
 GHIS0.67480.76190.28790.57660.48720.5826
 RVIS0.74400.82220.28360.60660.54290.5748
 HIS0.47590.50000.63270.67860.44930.5428
 IS0.70000.76670.41430.62730.58000.5767
 HIS (Imputed)0.38540.31400.72310.60000.44340.5479
 HIPred0.70730.73330.56820.63460.67570.7024
ASD 1
 EvoTol0.40160.24000.84780.63160.50650.4978
 GHIS0.74290.80850.30430.54290.60870.5185
 RVIS0.74680.80770.37780.60000.62960.6925
 HIS0.35630.20000.63160.33330.46150.4023
 IS0.51580.56600.40430.51720.45240.4621
 HIS (Imputed)0.36840.21740.74420.47620.47060.4426
 HIPred0.60490.66670.35420.50790.51520.4948
ASD 2
 EvoTol0.40150.29310.73080.54840.48100.4428
 GHIS0.67570.76470.22450.50650.47830.5646
 RVIS0.69050.75930.34000.55410.56670.6259
 HIS0.44900.41300.71430.65520.48080.5609
 IS0.62750.67240.46300.57350.56820.5923
 HIS (Imputed)0.37500.28570.72730.57140.44440.5483
 HIPred0.62110.66670.42590.53730.56100.5640

Note: NPV, negative predictive value; AUC, area under the curve.

Performance of methods used for predicting haploinsufficiency on known disease genes and mouse models Note: NPV, negative predictive value; AUC, area under the curve. Next, we tested these methods using a set of genes for which a heterozygous gene knockout causes ‘lethality’ phenotypes and seizures in mouse. After removing genes that were also present in our training data, we were left with 75 MGI Lethality and 90 MGI Seizure genes, which were matched using the same procedure as above. From our analysis, it would appear that HIPred outperforms all other methods across these datasets. Although the performance of HIPred appears to drop on the MGI Seizures datasets, the drop in performance is not as drastic as that observed with other methods. Using the number of associated publications in PubMed as a proxy of how well genes are studied, we tested whether the MGI dataset was enriched for less-studied genes compared with the OMIM datasets. Although the median number of publications was lower in the MGI datasets, we did not reach statistical significance using a Mann-Whitney U-test (P = 0.13 for MGI Seizures versus OMIM HI and P = 0.34 for MGI Lethality versus OMIM HI). Therefore, we also tested these methods on a set of candidate disease genes linked to autism (ASD1 and ASD2). These datasets were statistically enriched for less studied genes than the OMIM datasets (P = 0.02 for ASD1 versus OMIM HI and P = 0.01 ASD2 versus OMIM HI). After removing genes that were also present in our training data, we matched the remaining genes to to a random set of genes based on gene length as above. Our analysis shows that the performance of all methods drops significantly on these datasets, with RVIS performing best. The performance of HIPred is comparable to GHIS across the ASD datasets. However, it should be noted that we cannot be sure which ASD genes are casual (Steinberg ). Therefore, the results of this benchmark should be interpreted with some caution.

3.3 Rank correlation between methods

Following the above benchmarks, we tested the correlation in gene ranks between the methods (based on absolute Spearman’s rank correlation coefficient, see Table 3). Unsurprisingly, the highest correlation was observed between HIS and HIS (Imputed). Disregarding the HIS scores, rank correlations fall in the range 0.03–0.58, with correlations between EvoTol and all other methods being generally low. It appears that HIPred has a moderate correlation with all existing methods (coefficients range from 0.4994 to 0.5739, with the exception of EvoTol which yields a correlation coefficient of 0.0478).
Table 3

Spearman’s rank correlation between the methods

RVISISEvoTolHISHIS (imputed)GHISHIPred
RVIS1.0000
IS0.32931.0000
EvoTol0.04340.06751.0000
HIS0.32480.35340.05231.0000
HIS (Imputed)0.35120.38790.06090.99931.0000
GHIS0.56990.37830.03870.35980.36791.0000
HIPred0.49940.52500.04780.56520.57390.50311.0000
Spearman’s rank correlation between the methods

4 Discussion

In this study, we outlined HIPred, an integrative approach that combines genomic and evolutionary features with functional annotations from ENCODE and Roadmap Epigenomics to predict haploinsufficiency. We evaluated 3 approaches for data integration: integration at the data, kernel (MKL) and classifier (stacking) level; and observed improved performances over existing methods using all data integration techniques. In our experiments, we observed that MKL and stacking classifiers outperformed classifiers constructed for one type of data. However, we found that the most intuitive data integration technique, i.e. integration at the data level, outperformed other (more complex) data integration techniques. We observed comparable performances to existing methods using SVMs on the integrated data, but improved performances using a gradient boosted machine. The improved performance may be the result of the implicit feature selection performed in gradient boosted machines. Therefore, it may be possible to improve the performance of our MKL-based classifier using feature selection before data integration. However, our stacking classifier uses random forests (which are also tree-based methods similar to gradient boosted machines) for most feature groups and therefore performs some form of feature selection before data integration, so it is unclear how much benefit MKL would gain from feature selection. One main advantage to integration at the data level is the ability to capture the potential relationships between features across feature groups (which are missed using MKL and stacking based approaches). We benchmarked HIPred using several datasets and have shown that our method performs as well as, and in most cases, outperforms existing algorithms. An important issue to consider when comparing the performance of any prediction algorithm is the benchmark being used. Here, it is preferable to use blind datasets, i.e. data that have not been used to train the algorithm, to minimize any bias in the observed performance. Although we took care to reduce this bias in our results by performing an extensive cross-validation analysis and excluding genes from our benchmarks that were also present in our training data, this level of testing is not possible with alternative methods as it would require training each method using common data. Therefore, the performance of alternative methods may be inflated. Furthermore, it has been shown that most biological networks used in existing methods are effected by study bias (Steinberg ), i.e. well-studied genes are over-represented in these networks compared with less studied genes. As a result, existing algorithms may not generalize well to less studied genes. For example, the performance of most existing algorithms drops when predicting on the MGI datasets, which comprise less studied genes. In contrast, HIPred doesn’t appear to be affected by this study bias and outperforms existing methods on these datasets. Other important factors to consider when evaluating predictive methods are potential artifacts in performance arising due to gene similarity. Although we did not observe any gene similarities within our training data at the nucleotide sequence (up to 60% sequence similarity), we did not test for potential gene similarities at the protein sequence level. The advent of large genome sequencing consortia, such as the 1000 genomes project (The 1000 Genomes Project Consortium, 2012), NHLBI Exome Sequencing Project (ESP) and the ExAC (Lek ), creates an urgent need for unbiased haploinsufficiency prediction methods such as HIPred. Click here for additional data file.
  23 in total

1.  De novo mutations revealed by whole-exome sequencing are strongly associated with autism.

Authors:  Stephan J Sanders; Michael T Murtha; Abha R Gupta; John D Murdoch; Melanie J Raubeson; A Jeremy Willsey; A Gulhan Ercan-Sencicek; Nicholas M DiLullo; Neelroop N Parikshak; Jason L Stein; Michael F Walker; Gordon T Ober; Nicole A Teran; Youeun Song; Paul El-Fishawy; Ryan C Murtha; Murim Choi; John D Overton; Robert D Bjornson; Nicholas J Carriero; Kyle A Meyer; Kaya Bilguvar; Shrikant M Mane; Nenad Sestan; Richard P Lifton; Murat Günel; Kathryn Roeder; Daniel H Geschwind; Bernie Devlin; Matthew W State
Journal:  Nature       Date:  2012-04-04       Impact factor: 49.962

2.  Characterising and predicting haploinsufficiency in the human genome.

Authors:  Ni Huang; Insuk Lee; Edward M Marcotte; Matthew E Hurles
Journal:  PLoS Genet       Date:  2010-10-14       Impact factor: 5.917

3.  A systematic survey of loss-of-function variants in human protein-coding genes.

Authors:  Daniel G MacArthur; Suganthi Balasubramanian; Adam Frankish; Ni Huang; James Morris; Klaudia Walter; Luke Jostins; Lukas Habegger; Joseph K Pickrell; Stephen B Montgomery; Cornelis A Albers; Zhengdong D Zhang; Donald F Conrad; Gerton Lunter; Hancheng Zheng; Qasim Ayub; Mark A DePristo; Eric Banks; Min Hu; Robert E Handsaker; Jeffrey A Rosenfeld; Menachem Fromer; Mike Jin; Xinmeng Jasmine Mu; Ekta Khurana; Kai Ye; Mike Kay; Gary Ian Saunders; Marie-Marthe Suner; Toby Hunt; If H A Barnes; Clara Amid; Denise R Carvalho-Silva; Alexandra H Bignell; Catherine Snow; Bryndis Yngvadottir; Suzannah Bumpstead; David N Cooper; Yali Xue; Irene Gallego Romero; Jun Wang; Yingrui Li; Richard A Gibbs; Steven A McCarroll; Emmanouil T Dermitzakis; Jonathan K Pritchard; Jeffrey C Barrett; Jennifer Harrow; Matthew E Hurles; Mark B Gerstein; Chris Tyler-Smith
Journal:  Science       Date:  2012-02-17       Impact factor: 47.728

4.  Predicting the functional effect of amino acid substitutions and indels.

Authors:  Yongwook Choi; Gregory E Sims; Sean Murphy; Jason R Miller; Agnes P Chan
Journal:  PLoS One       Date:  2012-10-08       Impact factor: 3.240

5.  Genic intolerance to functional variation and the interpretation of personal genomes.

Authors:  Slavé Petrovski; Quanli Wang; Erin L Heinzen; Andrew S Allen; David B Goldstein
Journal:  PLoS Genet       Date:  2013-08-22       Impact factor: 5.917

6.  Analysis of protein-coding genetic variation in 60,706 humans.

Authors:  Monkol Lek; Konrad J Karczewski; Eric V Minikel; Kaitlin E Samocha; Eric Banks; Timothy Fennell; Anne H O'Donnell-Luria; James S Ware; Andrew J Hill; Beryl B Cummings; Taru Tukiainen; Daniel P Birnbaum; Jack A Kosmicki; Laramie E Duncan; Karol Estrada; Fengmei Zhao; James Zou; Emma Pierce-Hoffman; Joanne Berghout; David N Cooper; Nicole Deflaux; Mark DePristo; Ron Do; Jason Flannick; Menachem Fromer; Laura Gauthier; Jackie Goldstein; Namrata Gupta; Daniel Howrigan; Adam Kiezun; Mitja I Kurki; Ami Levy Moonshine; Pradeep Natarajan; Lorena Orozco; Gina M Peloso; Ryan Poplin; Manuel A Rivas; Valentin Ruano-Rubio; Samuel A Rose; Douglas M Ruderfer; Khalid Shakir; Peter D Stenson; Christine Stevens; Brett P Thomas; Grace Tiao; Maria T Tusie-Luna; Ben Weisburd; Hong-Hee Won; Dongmei Yu; David M Altshuler; Diego Ardissino; Michael Boehnke; John Danesh; Stacey Donnelly; Roberto Elosua; Jose C Florez; Stacey B Gabriel; Gad Getz; Stephen J Glatt; Christina M Hultman; Sekar Kathiresan; Markku Laakso; Steven McCarroll; Mark I McCarthy; Dermot McGovern; Ruth McPherson; Benjamin M Neale; Aarno Palotie; Shaun M Purcell; Danish Saleheen; Jeremiah M Scharf; Pamela Sklar; Patrick F Sullivan; Jaakko Tuomilehto; Ming T Tsuang; Hugh C Watkins; James G Wilson; Mark J Daly; Daniel G MacArthur
Journal:  Nature       Date:  2016-08-18       Impact factor: 49.962

7.  An integrated encyclopedia of DNA elements in the human genome.

Authors: 
Journal:  Nature       Date:  2012-09-06       Impact factor: 49.962

8.  An integrated map of genetic variation from 1,092 human genomes.

Authors:  Goncalo R Abecasis; Adam Auton; Lisa D Brooks; Mark A DePristo; Richard M Durbin; Robert E Handsaker; Hyun Min Kang; Gabor T Marth; Gil A McVean
Journal:  Nature       Date:  2012-11-01       Impact factor: 49.962

9.  Predicting the functional, molecular, and phenotypic consequences of amino acid substitutions using hidden Markov models.

Authors:  Hashem A Shihab; Julian Gough; David N Cooper; Peter D Stenson; Gary L A Barker; Keith J Edwards; Ian N M Day; Tom R Gaunt
Journal:  Hum Mutat       Date:  2012-11-02       Impact factor: 4.878

10.  Integrative analysis of 111 reference human epigenomes.

Authors:  Anshul Kundaje; Wouter Meuleman; Jason Ernst; Misha Bilenky; Angela Yen; Alireza Heravi-Moussavi; Pouya Kheradpour; Zhizhuo Zhang; Jianrong Wang; Michael J Ziller; Viren Amin; John W Whitaker; Matthew D Schultz; Lucas D Ward; Abhishek Sarkar; Gerald Quon; Richard S Sandstrom; Matthew L Eaton; Yi-Chieh Wu; Andreas R Pfenning; Xinchen Wang; Melina Claussnitzer; Yaping Liu; Cristian Coarfa; R Alan Harris; Noam Shoresh; Charles B Epstein; Elizabeta Gjoneska; Danny Leung; Wei Xie; R David Hawkins; Ryan Lister; Chibo Hong; Philippe Gascard; Andrew J Mungall; Richard Moore; Eric Chuah; Angela Tam; Theresa K Canfield; R Scott Hansen; Rajinder Kaul; Peter J Sabo; Mukul S Bansal; Annaick Carles; Jesse R Dixon; Kai-How Farh; Soheil Feizi; Rosa Karlic; Ah-Ram Kim; Ashwinikumar Kulkarni; Daofeng Li; Rebecca Lowdon; GiNell Elliott; Tim R Mercer; Shane J Neph; Vitor Onuchic; Paz Polak; Nisha Rajagopal; Pradipta Ray; Richard C Sallari; Kyle T Siebenthall; Nicholas A Sinnott-Armstrong; Michael Stevens; Robert E Thurman; Jie Wu; Bo Zhang; Xin Zhou; Arthur E Beaudet; Laurie A Boyer; Philip L De Jager; Peggy J Farnham; Susan J Fisher; David Haussler; Steven J M Jones; Wei Li; Marco A Marra; Michael T McManus; Shamil Sunyaev; James A Thomson; Thea D Tlsty; Li-Huei Tsai; Wei Wang; Robert A Waterland; Michael Q Zhang; Lisa H Chadwick; Bradley E Bernstein; Joseph F Costello; Joseph R Ecker; Martin Hirst; Alexander Meissner; Aleksandar Milosavljevic; Bing Ren; John A Stamatoyannopoulos; Ting Wang; Manolis Kellis
Journal:  Nature       Date:  2015-02-19       Impact factor: 69.504

View more
  13 in total

Review 1.  Human gene essentiality.

Authors:  István Bartha; Julia di Iulio; J Craig Venter; Amalio Telenti
Journal:  Nat Rev Genet       Date:  2017-10-30       Impact factor: 53.242

Review 2.  Epigenetic and Cellular Diversity in the Brain through Allele-Specific Effects.

Authors:  Wei-Chao Huang; Kathleen Bennett; Christopher Gregg
Journal:  Trends Neurosci       Date:  2018-08-08       Impact factor: 13.837

3.  Promoter CpG Density Predicts Downstream Gene Loss-of-Function Intolerance.

Authors:  Leandros Boukas; Hans T Bjornsson; Kasper D Hansen
Journal:  Am J Hum Genet       Date:  2020-08-14       Impact factor: 11.025

4.  Translational enhancement by base editing of the Kozak sequence rescues haploinsufficiency.

Authors:  Chiara Ambrosini; Eliana Destefanis; Eyemen Kheir; Francesca Broso; Federica Alessandrini; Sara Longhi; Nicolò Battisti; Isabella Pesce; Erik Dassi; Gianluca Petris; Anna Cereseto; Alessandro Quattrone
Journal:  Nucleic Acids Res       Date:  2022-10-14       Impact factor: 19.160

Review 5.  Machine learning for epigenetics and future medical applications.

Authors:  Lawrence B Holder; M Muksitul Haque; Michael K Skinner
Journal:  Epigenetics       Date:  2017-05-19       Impact factor: 4.528

Review 6.  Of numbers and movement - understanding transcription factor pathogenesis by advanced microscopy.

Authors:  Julia M T Auer; Jack J Stoddart; Ioannis Christodoulou; Ana Lima; Kassiani Skouloudaki; Hildegard N Hall; Vladana Vukojević; Dimitrios K Papadopoulos
Journal:  Dis Model Mech       Date:  2020-12-29       Impact factor: 5.758

7.  FATHMM-XF: accurate prediction of pathogenic point mutations via extended features.

Authors:  Mark F Rogers; Hashem A Shihab; Matthew Mort; David N Cooper; Tom R Gaunt; Colin Campbell
Journal:  Bioinformatics       Date:  2018-02-01       Impact factor: 6.937

8.  Transcriptome and translatome co-evolution in mammals.

Authors:  Zhong-Yi Wang; Evgeny Leushkin; Angélica Liechti; Svetlana Ovchinnikova; Katharina Mößinger; Thoomke Brüning; Coralie Rummel; Frank Grützner; Margarida Cardoso-Moreira; Peggy Janich; David Gatfield; Boubou Diagouraga; Bernard de Massy; Mark E Gill; Antoine H F M Peters; Simon Anders; Henrik Kaessmann
Journal:  Nature       Date:  2020-11-11       Impact factor: 49.962

9.  Mutation severity spectrum of rare alleles in the human genome is predictive of disease type.

Authors:  Jimin Pei; Lisa N Kinch; Zbyszek Otwinowski; Nick V Grishin
Journal:  PLoS Comput Biol       Date:  2020-05-15       Impact factor: 4.475

10.  Characterization of Human Dosage-Sensitive Transcription Factor Genes.

Authors:  Zhihua Ni; Xiao-Yu Zhou; Sidra Aslam; Deng-Ke Niu
Journal:  Front Genet       Date:  2019-12-04       Impact factor: 4.599

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.