| Literature DB >> 20684785 |
Abstract
BACKGROUND: To further understand the implementation of hyperparameters re-estimation technique in Bayesian hierarchical model, we added two more prior assumptions over the weight in BayesPI, namely Laplace prior and Cauchy prior, by using the evidence approximation method. In addition, we divided hyperparameter (regularization constants alpha of the model) into multiple distinct classes based on either the structure of the neural networks or the property of the weights.Entities:
Mesh:
Substances:
Year: 2010 PMID: 20684785 PMCID: PMC2921412 DOI: 10.1186/1471-2105-11-412
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Figure 1Performance comparisons from simulated ChIP-chip datasets. The upper panel of the figure shows the box plots of the distribution of motif similarity scores across 15 different weight prior configurations. The lower panel of the figures shows the box plots of the distribution of CPU hours used by 15 prior assumptions over the weights. Here, the red line represents Gaussian prior assumption to the weights (e.g. G1, G2, G3, G4, and G5), the blue line represents Laplace prior approximation over the weights (e.g. L1, L2, L3, L4, and L5), and the black line indicates Cauchy priors to the weights (C1, C2, C3, C4, and C5), in which the numerical values 1, 2, 3, 4, and 5 represent regularization constant α with one, two, three, four, and greater than five classes, respectively.
Figure 2Performance comparisons from real ChIP-chip datasets. The upper panel of the figure shows the box plots of the distribution of motif similarity scores across six different weight prior configurations. The lower panel of the figure shows the box plots of the distribution of CPU hours used by six prior assumptions over the weights. Here, the red line represents the prior assumptions over the weights without inclusion of the nucleosome information, and the blue line indicates the prior assumptions over the weights with the inclusion of the nucleosome information. G4, L4, and C4 indicates Gaussian prior, Laplace prior, and Cauchy prior assumptions to the weights with four classes of regularization constants α, respectively.
Comparing motif similarity scores of nine yeast TFs from four different calculations.
| TF Name (consensus sequence length) | Activated in stress conditions | BayesPI - Gaussian prior | BayesPI - Laplace prior | BayesPI - Cauchy prior | MatrixREDUCE |
|---|---|---|---|---|---|
| ACE2 (6) | No | 0.89 | 0.95 | 0.96 | 0.90 |
| 0.93 | |||||
| SWI4 (7) | No | 0.96 | 0.94 | 0.94 | 0.95 |
| YAP1 (7) | Yes[ | 0.93 | 0.92 | 0.92 | 0.93 |
| INO4 (8) | No | 0.90 | 0.92 | 0.94 | 0.97 |
| 0.86 | 0.87 | 0.86 | |||
| FHL1 (10) | No | 0.95 | 0.95 | 0.93 | 0.88 |
For the nine yeast TFs, the ChIP-chip datasets were obtained from [7]; the regularization constants α in BayesPI were divided into four classes; MatrixREDUCE program was downloaded from the publication [8] and its default parameters were used in the present study. Here, the motif similarity scores greater than 0.85 represents a good match between the prediction and the SGD consensus sequences [1]. Poor predictions are marked by bold text. NA indicates that no results are available owing to the program reason. All the programs were applied on the same datasets and were run under a PC cluster (a dual-core CPU SUN X6220 blade node with 16 GB of RAM).
Figure 3Performance comparisons from human ChIP-Seq datasets. The upper panel of the figure shows the box plots of the distribution of motif similarity scores across three different weight prior configurations. The lower panel of the figure shows the box plots of the distribution of CPU hours used by three prior assumptions over the weights. The red line represents Gaussian prior assumption over the weights, the blue line indicates the Laplace prior assumption over the weights, and the black line represents Cauchy prior assumption to the weights. Here, there are four distinct weight classes in the regularization constants α.