Bayesian learning of inverted Dirichlet mixtures for SVM kernels generation

被引:38
作者
Bdiri, Taoufik [1 ]
Bouguila, Nizar [2 ]
机构
[1] Concordia Univ, Elect & Comp Engn Dept ECE, Montreal, PQ H3G 1T7, Canada
[2] Concordia Univ, Concordia Inst Informat Syst Engn CIISE, Montreal, PQ H3G 1T7, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Mixture models; SVM; Hybrid models; Inverted Dirichlet; Bayesian inference; Bayes factor; Model selection; Gibbs sampling; Kernels; Object detection; Image databases; NORMALIZING CONSTANTS; MONTE-CARLO; FIT DATA; MODEL; CONVERGENCE; APPROXIMATIONS; SIMULATION; CHOICE; RATIOS; RATES;
D O I
10.1007/s00521-012-1094-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We describe approaches for positive data modeling and classification using both finite inverted Dirichlet mixture models and support vector machines (SVMs). Inverted Dirichlet mixture models are used to tackle an outstanding challenge in SVMs namely the generation of accurate kernels. The kernels generation approaches, grounded on ideas from information theory that we consider, allow the incorporation of data structure and its structural constraints. Inverted Dirichlet mixture models are learned within a principled Bayesian framework using both Gibbs sampler and Metropolis-Hastings for parameter estimation and Bayes factor for model selection (i.e., determining the number of mixture's components). Our Bayesian learning approach uses priors, which we derive by showing that the inverted Dirichlet distribution belongs to the family of exponential distributions, over the model parameters, and then combines these priors with information from the data to build posterior distributions. We illustrate the merits and the effectiveness of the proposed method with two real-world challenging applications namely object detection and visual scenes analysis and classification.
引用
收藏
页码:1443 / 1458
页数:16
相关论文
共 85 条
[61]   Fitting full-information item factor models and an empirical investigation of bridge sampling [J].
Meng, XL ;
Schilling, S .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1996, 91 (435) :1254-1267
[62]   A global optimization technique for statistical classifier design [J].
Miller, D ;
Rao, AV ;
Rose, K ;
Gersho, A .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 1996, 44 (12) :3108-3122
[63]   BAYESIAN METHODS IN PRACTICE - EXPERIENCES IN THE PHARMACEUTICAL-INDUSTRY [J].
RACINE, A ;
GRIEVE, AP ;
FLUHLER, H ;
SMITH, AFM .
APPLIED STATISTICS-JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES C, 1986, 35 (02) :93-150
[64]   Approximate Bayes factors and accounting for model uncertainty in generalised linear models [J].
Raftery, AE .
BIOMETRIKA, 1996, 83 (02) :251-266
[65]  
Renyi Alfred., 1960, 4th Berkeley Symposium on Mathematics, Statistics and Probability, V547, P547
[66]   MODELING BY SHORTEST DATA DESCRIPTION [J].
RISSANEN, J .
AUTOMATICA, 1978, 14 (05) :465-471
[67]   Optimal scaling of discrete approximations to Langevin diffusions [J].
Roberts, GO ;
Rosenthal, JS .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 1998, 60 :255-268
[68]   Bounds on regeneration times and convergence rates for Markov chains [J].
Roberts, GO ;
Tweedie, RL .
STOCHASTIC PROCESSES AND THEIR APPLICATIONS, 1999, 80 (02) :211-229
[69]  
ROBERTS GO, 1994, J ROY STAT SOC B MET, V56, P377
[70]  
Rosenthal J. S., 1993, Annals of Applied Probability, V3, P819, DOI [DOI 10.1214/AOAP/1177005366, https://doi.org/10.1214/aoap/1177005366]