Partially collapsed parallel Gibbs sampler for Dirichlet process mixture models

被引:7
作者
Yerebakan, Halid Ziya [1 ]
Dundar, Murat [1 ]
机构
[1] IUPUI Indianapolis, Comp & Informat Sci, Indianapolis, IN 46202 USA
基金
美国国家科学基金会;
关键词
Dirichlet process; Mixture models; Clustering; Markov chain Monte Carlo;
D O I
10.1016/j.patrec.2017.03.009
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dirichlet Process (DP) is commonly used as a non-parametric prior on mixture models. It has adaptive model selection capability which is useful in clustering applications. Although exact inference is not tractable for this prior, Markov Chain Monte Carlo (MCMC) samplers have been used to approximate the target posterior distribution. These samplers often do not scale well. Thus, recent studies focused on improving run-time efficiency through parallelization. In this paper, we introduce a new sampling method for DP by combining Chinese Restaurant Process (CRP) with stick-breaking construction allowing for parallelization through conditional independence at the data point level. Stick breaking part uses an uncollapsed sampler providing a high level of parallelization while the CRP part uses collapsed sampler allowing more accurate clustering. We show that this partially collapsed Gibbs sampler has significant advantages over the collapsed-only version in terms of scalability. We also provide results on real-world data sets that favorably compares the proposed inference algorithm against a recently introduced parallel Dirichlet Process samplers in terms of F1 scores while maintaining a comparable run-time performance. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:22 / 27
页数:6
相关论文
共 20 条
[1]  
Aldous D., 1983, ECOLE DETE PROBABILI, P1
[2]  
[Anonymous], TECHNICAL REPORT
[3]  
[Anonymous], 2013, Advances in Neural Information Processing Systems
[4]  
Blei D.M., 2004, Advances in Neural Information Processing Systems, P2003
[5]   Efficient Classification-Based Relabeling in Mixture Models [J].
Cron, Andrew J. ;
West, Mike .
AMERICAN STATISTICIAN, 2011, 65 (01) :16-20
[6]  
Dahl D.B., 2003, TECH REP, V1
[7]   BAYESIAN ANALYSIS OF SOME NONPARAMETRIC PROBLEMS [J].
FERGUSON, TS .
ANNALS OF STATISTICS, 1973, 1 (02) :209-230
[8]  
Gal Y., 2013, WORKSH BIG LEARN NIP
[9]  
Ge H, 2015, PR MACH LEARN RES, V37, P2276
[10]  
Görür D, 2010, J COMPUT SCI TECH-CH, V25, P653, DOI [10.1007/s11390-010-1051-1, 10.1007/s11390-010-9355-8]