| Literature DB >> 18447951 |
Alexander Churbanov1, Stephen Winters-Hilt.
Abstract
BACKGROUND: The Baum-Welch learning procedure for Hidden Markov Models (HMMs) provides a powerful tool for tailoring HMM topologies to data for use in knowledge discovery and clustering. A linear memory procedure recently proposed by Miklós, I. and Meyer, I.M. describes a memory sparse version of the Baum-Welch algorithm with modifications to the original probabilistic table topologies to make memory use independent of sequence length (and linearly dependent on state number). The original description of the technique has some errors that we amend. We then compare the corrected implementation on a variety of data sets with conventional and checkpointing implementations.Entities:
Mesh:
Substances:
Year: 2008 PMID: 18447951 PMCID: PMC2430973 DOI: 10.1186/1471-2105-9-224
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
The computational expense of different algorithm implementations running on HMM.
| Algorith | Canonical | Checkpointing | Linear | |||
| Viterbi | Time | Time | Time | |||
| Space | Space | Space | ||||
| Baum-Welch | Time | Time | Time | |||
| Space | Space | Space | ||||
The Viterbi decoding, forward and backward procedures.
| Forward procedure | Backward procedure | Viterbi algorithm |
| • Initially | ||
| • Initially | • Initially | • |
| • | • | • Finally |
| • Finally | • Finally |
The maximization step in HMM learning. states.
| Initial probability estimate | Transition probability estimate | Emission parameters estimate |
| • Gaussian emission | ||
| • Discrete emission |
Figure 1Time trellis for simple model where possible emissions of 0 and 1 are shown above and below trellis. Probabilities of emissions that happen after each transition are shown in bold and transitions of interest taken at certain time-point are underlined.
Figure 2The linear memory implementation of Baum-Welch learning algorithm for HMM. This algorithm takes set of HMM parameters λ and sequence of symbols O. Expected HMM parameters are calculated according to formulas [see Subsection Parameters update].
The scoring functions for discrete and continuous emissions.
| Discrete emission | Continuous Gaussian emission |
| Score ( | Score ( |
| | |
| | |
| |
Figure 3Viterbi algorithm implementation with linked list. Here is reference to the previous node.
Figure 4Explicit DHMM topology. Here the first aggregate state emits 0 with probability 0.75 and 1 with probability 0.25 and the second aggregate state emits 0 with probability 0.25 and 1 with probability 0.75. Duration histograms are shown for each aggregate state.
Figure 5Explicit Duration HMM trellis for the observation string shown below. The most likely sequence of states for the observation string shown below is shaded. The lightly grayed states will be deallocated by garbage collector.
Figure 6Spike detection loop topology.
Figure 7Trellis for loopy topology used for spikes detection where shallow spike (states 1–6) and deep spike (states 7–17) are consequently decoded. The most likely sequence of states for the sequence of observed ionic flow current blockades (in pA) shown below is shaded. The lightly grayed states will be deallocated by garbage collector.
Figure 8In subfigures 8(a) – 8(c) performance of Baum-Welch used on DHMM topology with two aggregate states of various maximum duration In subfigures 8(d) – 8(f) performance of Baum-Welch algorithm used on spike topology for various ionic flow durations is shown.
Memory use for Viterbi decoding on spike topology with loop sizes 6 and 11.
| Ionic flow samples | Ratio of number of referenced links to sequence size |
| 819 | 1.1173 |
| 10,319 | 1.0084 |
| 26,233 | 1.0042 |
| 51,233 | 1.0017 |
| 101,233 | 1.0015 |
| 151,232 | 1.0007 |
Memory use for Viterbi decoding on explicit DHMM with D = 60 and two aggregate
| Observation sequence size | Ratio of number of referenced links to sequence size |
| 1,000 | 2.565 |
| 10,000 | 1.134 |
| 50,000 | 1.032 |
| 100,000 | 1.017 |
| 200,000 | 1.007 |