Bayesian Inference with Posterior Regularization and Applications to Infinite Latent SVMs

被引:0
作者
Zhu, Jun [1 ]
Chen, Ning [1 ]
Xing, Eric P. [2 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, Tsinghua Natl Lab Informat Sci & Technol, State Key Lab Intelligent Technol & Syst, Beijing 100084, Peoples R China
[2] Carnegie Mellon Univ, Sch Comp Sci, Pittsburgh, PA 15213 USA
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Bayesian inference; posterior regularization; Bayesian nonparametrics; large-margin learning; classification; multi-task learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Existing Bayesian models, especially nonparametric Bayesian methods, rely on specially conceived priors to incorporate domain knowledge for discovering improved latent representations. While priors affect posterior distributions through Bayes' rule, imposing posterior regularization is arguably more direct and in some cases more natural and general. In this paper, we present regularized Bayesian inference (RegBayes), a novel computational framework that performs posterior inference with a regularization term on the desired post-data posterior distribution under an information theoretical formulation. RegBayes is more flexible than the procedure that elicits expert knowledge via priors, and it covers both directed Bayesian networks and undirected Markov networks. When the regularization is induced from a linear operator on the posterior distributions, such as the expectation operator, we present a general convex-analysis theorem to characterize the solution of RegBayes. Furthermore, we present two concrete examples of RegBayes, infinite latent support vector machines (iLSVM) and multi-task infinite latent support vector machines (MT-iLSVM), which explore the large-margin idea in combination with a nonparametric Bayesian model for discovering predictive latent features for classification and multi-task learning, respectively. We present efficient inference methods and report empirical studies on several benchmark data sets, which appear to demonstrate the merits inherited from both large-margin learning and Bayesian nonparametrics. Such results contribute to push forward the interface between these two important subfields, which have been largely treated as isolated in the community.
引用
收藏
页码:1799 / 1847
页数:49
相关论文
共 74 条
[1]   Unifying divergence minimization and statistical inference via convex duality [J].
Altun, Yasemin ;
Smola, Alex .
LEARNING THEORY, PROCEEDINGS, 2006, 4005 :139-153
[2]  
Ando RK, 2005, J MACH LEARN RES, V6, P1817
[3]  
[Anonymous], 2009, P 26 ANN INT C MACH
[4]  
[Anonymous], 2008, Pushing the limits of contemporary statistics: contributions in honor of Jayanta K. Ghosh, DOI 10.1214/074921708000000138
[5]  
[Anonymous], 2009, Advances in neural information processing systems
[6]  
[Anonymous], 2003, BAYESIAN NONPARAMETR
[7]  
[Anonymous], 2007, Multi-Task Feature Learning, DOI DOI 10.7551/MITPRESS/7503.003.0010
[8]  
[Anonymous], THESIS U CAMBRIDGE
[9]  
[Anonymous], 2007, Artificial intelligence and statistics
[10]  
[Anonymous], 2005, TECHNICAL REPORT