Learning Topic Models by Belief Propagation

被引:34
作者
Zeng, Jia [1 ]
Cheung, William K. [2 ]
Liu, Jiming [2 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou 215006, Peoples R China
[2] Hong Kong Baptist Univ, Dept Comp Sci, Kowloon Tong, Hong Kong, Peoples R China
关键词
Latent Dirichlet allocation; topic models; belief propagation; message passing; factor graph; Bayesian networks; Markov random fields; hierarchical Bayesian models; Gibbs sampling; variational Bayes; EM;
D O I
10.1109/TPAMI.2012.185
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Latent Dirichlet allocation (LDA) is an important hierarchical Bayesian model for probabilistic topic modeling, which attracts worldwide interest and touches on many important applications in text mining, computer vision and computational biology. This paper represents the collapsed LDA as a factor graph, which enables the classic loopy belief propagation (BP) algorithm for approximate inference and parameter estimation. Although two commonly used approximate inference methods, such as variational Bayes (VB) and collapsed Gibbs sampling (GS), have gained great success in learning LDA, the proposed BP is competitive in both speed and accuracy, as validated by encouraging experimental results on four large-scale document datasets. Furthermore, the BP algorithm has the potential to become a generic scheme for learning variants of LDA-based topic models in the collapsed space. To this end, we show how to learn two typical variants of LDA-based topic models, such as author-topic models (ATM) and relational topic models (RTM), using BP based on the factor graph representations.
引用
收藏
页码:1121 / 1134
页数:14
相关论文
共 39 条
[11]  
Eisenstein J., 2010, TECHNICAL REPORT
[12]   Clustering by passing messages between data points [J].
Frey, Brendan J. ;
Dueck, Delbert .
SCIENCE, 2007, 315 (5814) :972-976
[13]   Finding scientific topics [J].
Griffiths, TL ;
Steyvers, M .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2004, 101 :5228-5235
[14]  
Hoffman MD, 2010, ADV NEURAL INFORM PR, V1, P856, DOI DOI 10.5555/2997189.2997285
[15]   Unsupervised learning by probabilistic latent semantic analysis [J].
Hofmann, T .
MACHINE LEARNING, 2001, 42 (1-2) :177-196
[16]  
Jia Zeng, 2010, Proceedings 2010 IEEE/ACM International Conference on Web Intelligence-Intelligent Agent Technology (WI-IAT), P366, DOI 10.1109/WI-IAT.2010.20
[17]   Factor graphs and the sum-product algorithm [J].
Kschischang, FR ;
Frey, BJ ;
Loeliger, HA .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2001, 47 (02) :498-519
[18]   Learning the parts of objects by non-negative matrix factorization [J].
Lee, DD ;
Seung, HS .
NATURE, 1999, 401 (6755) :788-791
[19]   Automating the construction of internet portals with machine learning [J].
McCallum, AK ;
Nigam, K ;
Rennie, J ;
Seymore, K .
INFORMATION RETRIEVAL, 2000, 3 (02) :127-163
[20]  
Minka T.P., 2002, Uncertainty in Artificial Intelligence, P352