| Literature DB >> 20175896 |
Jens Keilwagen1, Jan Grau, Stefan Posch, Marc Strickert, Ivo Grosse.
Abstract
BACKGROUND: The recognition of functional binding sites in genomic DNA remains one of the fundamental challenges of genome research. During the last decades, a plethora of different and well-adapted models has been developed, but only little attention has been payed to the development of different and similarly well-adapted learning principles. Only recently it was noticed that discriminative learning principles can be superior over generative ones in diverse bioinformatics applications, too.Entities:
Mesh:
Substances:
Year: 2010 PMID: 20175896 PMCID: PMC2848239 DOI: 10.1186/1471-2105-11-98
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Learning principles
| prior knowledge | |||
|---|---|---|---|
| non Bayesian | Bayesian | ||
| ML | MAP | ||
| GDT | PGDT | ||
| MCL | MSP | ||
The table shows six established learning principles that can be grouped by their objective as being generative, hybrid, or discriminative and utilization of prior knowledge with the two possibilities non Bayesian and Bayesian. The four elementary learning principles are the generative, non Bayesian maximum likelihood (ML) learning principle, the generative, Bayesian maximum a posteriori (MAP) learning principle, the discriminative, non Bayesian maximum conditional likelihood (MCL) learning principle, and the discriminative, Bayesian maximum supervised posterior (MSP) learning principle. The hybrid learning principles which interpolate between generative and discriminative learning principles are the non Bayesian generative-discriminative trade-off (GDT) learning principle and the penalized generative-discriminative trade-off (PGDT) learning principle.
Figure 1Illustration of the unified generative-discriminative learning principle. The plots show a projection of the simplex onto the (β0, β1)-plane and the corresponding learning principles for the specific weights encoded by colors. Figure 1(a) shows the general interpretation of the simplex where the points (0, 1), (0, 0.5), (1, 0), and (0.5, 0) refer to the ML, MAP, MCL, and MSP learning principle, respectively, while the lines β1 = 1 - β0 and β1 = 0.5 - β0 refer to the GDT and PGDT learning principle, respectively. Figure 1(b) shows the interpretation of the unified generative-discriminative learning principle for a conjugate prior that satisfies the condition of equation (11). In this case, each point on the abscissa (β0-axis) and ordinate (β1-axis) refers to the MSP and MAP learning principle, respectively, using the prior in a weighted version . The simplex colored in gray corresponds to the MSP learning principle using the weighted posterior as prior for the parameter vector .
Figure 2Performance of the unified generative-discriminative learning principle for four data sets. We perform a 1,000-fold stratified hold-out sampling procedure for the four data sets, record for different values of the mean sensitivity for a fixed specificity of 99.9%, and plot the mean sensitivities on the simplex in analogy to Figure 1. Yellow indicates the highest sensitivity, red indicates the lowest sensitivity, and the gray contour lines of each subfigure indicate multiples of the standard error of the maximum sensitivity.
Results for four data sets
| AR/GR/PR | GATA | NF- | Thyroid | |
|---|---|---|---|---|
| ML | 54.7 | 77.0 | 81.6 | 51.3 |
| MCL | 55.2 | 73.2 | 76.5 | 50.0 |
| MAP | 55.1 | 77.0 | 81.6 | 51.3 |
| MSP | 56.9 | 77.0 | 79.6 | 50.4 |
| Unified | ||||
Summary the results of Figure 2 for the 4 data sets containing the highest sensitivity for the ML, the MCL, the MAP, the MSP, and the unified generative-discriminative learning principle. For the MAP, the MSP, and the unified generative-discriminative learning principle, we present the best results form the simplex which correspond to one of these learning principles (see Figure 1b). For each data set, the highest sensitivity is displayed in bold.