A latent variable Gaussian process model with Pitman-Yor process priors for multiclass classification

被引:0
作者
Chatzis, Sotirios P.
机构
[1] Department of Electrical Engineering, Computer Engineering, And Informatics, Cyprus University of Technology
关键词
Gaussian process; Pitman-Yor process; Mixture model; DISTRIBUTIONS; INFERENCE; MIXTURES;
D O I
10.1016/j.neucom.2013.04.029
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Gaussian processes (GPs) constitute one of the most important Bayesian machine learning approaches. Several researchers have considered postulating mixtures of Gaussian processes as a means of dealing with non-stationary covariance functions, discontinuities, multi-modality, and overlapping output signals. In existing works, mixtures of Gaussian processes are based on the introduction of a gating function defined over the space of model input variables. This way, each postulated mixture component Gaussian process is effectively restricted in a limited subset of the input space. Additionally, the applicability of these models is limited to regression tasks. In this paper, for the first time in the literature, we devise a Gaussian process mixture model especially suitable for multiclass classification applications: We consider a GP classification scheme the prior distribution of which is a fully generative nonparametric Bayesian model with power-law behavior, generating Gaussian processes over the whole input space of the learned task. We provide an efficient algorithm for model inference, based on the variational Bayesian framework, and exhibit its efficacy using benchmark and real-world classification datasets. (c) 2013 Elsevier B.V. All rights reserved.
引用
收藏
页码:482 / 489
页数:8
相关论文
共 37 条
[1]  
[Anonymous], 2006, Advances in Neural Information Processing Systems
[2]   MIXTURES OF DIRICHLET PROCESSES WITH APPLICATIONS TO BAYESIAN NONPARAMETRIC PROBLEMS [J].
ANTONIAK, CE .
ANNALS OF STATISTICS, 1974, 2 (06) :1152-1174
[3]   Audio thumbnailing of popular music using chroma-based representations [J].
Bartsch, MA ;
Wakefield, GH .
IEEE TRANSACTIONS ON MULTIMEDIA, 2005, 7 (01) :96-104
[4]  
Bertin-Mahieux T., P 12 INT C MUS INF R
[5]  
Bishop CM., 1995, NEURAL NETWORKS PATT
[6]   FERGUSON DISTRIBUTIONS VIA POLYA URN SCHEMES [J].
BLACKWELL, D ;
MACQUEEN, JB .
ANNALS OF STATISTICS, 1973, 1 (02) :353-355
[7]  
Blei D. M., 2004, International Conference on Machine Learning, P12, DOI DOI 10.1145/1015330.1015439
[8]   Variational Inference for Dirichlet Process Mixtures [J].
Blei, David M. ;
Jordan, Michael I. .
BAYESIAN ANALYSIS, 2006, 1 (01) :121-143
[9]   LIBSVM: A Library for Support Vector Machines [J].
Chang, Chih-Chung ;
Lin, Chih-Jen .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2011, 2 (03)
[10]  
Chatzis S. P., 2011, P ICTAI 11