| Literature DB >> 20871805 |
David Andrzejewski1, Xiaojin Zhu, Mark Craven.
Abstract
Users of topic modeling methods often have knowledge about the composition of words that should have high or low probability in various topics. We incorporate such domain knowledge using a novel Dirichlet Forest prior in a Latent Dirichlet Allocation framework. The prior is a mixture of Dirichlet tree distributions with special structures. We present its construction, and inference via collapsed Gibbs sampling. Experiments on synthetic and real datasets demonstrate our model's ability to follow and generalize beyond user-specified domain knowledge.Entities:
Year: 2009 PMID: 20871805 PMCID: PMC2943854 DOI: 10.1145/1553374.1553378
Source DB: PubMed Journal: Proc Int Conf Mach Learn