Novel mixture allocation models for topic learning

被引:0
作者
Maanicshah, Kamal [1 ,3 ]
Amayri, Manar [2 ]
Bouguila, Nizar [1 ]
机构
[1] Concordia Univ, Concordia Inst Informat & Syst Engn, Quebec City, PQ, Canada
[2] Grenoble Inst Technol, G SCOP Lab, Grenoble, France
[3] Concordia Univ, Concordia Inst Informat Syst Engn, Montreal, PQ, Canada
关键词
Beta-Liouville allocation; generalized Dirichlet allocation; mixture models; online learning; topic models; DIRICHLET; SELECTION; CLASSIFICATION; INFERENCE;
D O I
10.1111/coin.12641
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Latent Dirichlet allocation (LDA) is one of the major models used for topic modelling. A number of models have been proposed extending the basic LDA model. There has also been interesting research to replace the Dirichlet prior of LDA with other pliable distributions like generalized Dirichlet, Beta-Liouville and so forth. Owing to the proven efficiency of using generalized Dirichlet (GD) and Beta-Liouville (BL) priors in topic models, we use these versions of topic models in our paper. Furthermore, to enhance the support of respective topics, we integrate mixture components which gives rise to generalized Dirichlet mixture allocation and Beta-Liouville mixture allocation models respectively. In order to improve the modelling capabilities, we use variational inference method for estimating the parameters. Additionally, we also introduce an online variational approach to cater to specific applications involving streaming data. We evaluate our models based on its performance on applications related to text classification, image categorization and genome sequence classification using a supervised approach where the labels are used as an observed variable within the model.
引用
收藏
页数:29
相关论文
共 32 条
[1]  
[Anonymous], 2011, P 49 ANN M ASS COMP
[2]   A latent Beta-Liouville allocation model [J].
Bakhtiari, Ali Shojaee ;
Bouguila, Nizar .
EXPERT SYSTEMS WITH APPLICATIONS, 2016, 45 :260-272
[3]   A variational Bayes model for count data learning and classification [J].
Bakhtiari, Ali Shojaee ;
Bouguila, Nizar .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2014, 35 :176-186
[4]   Variational Bayesian inference for infinite generalized inverted Dirichlet mixtures with feature selection and its application to clustering [J].
Bdiri, Taoufik ;
Bouguila, Nizar ;
Ziou, Djemel .
APPLIED INTELLIGENCE, 2016, 44 (03) :507-525
[5]   Dirichlet Mixture Allocation for Multiclass Document Collections Modeling [J].
Bian, Wei ;
Tao, Dacheng .
2009 9TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING, 2009, :711-715
[6]  
Blei D., 2006, Advances in Neural Information Processing Systems, V18, P147
[7]   Variational Inference: A Review for Statisticians [J].
Blei, David M. ;
Kucukelbir, Alp ;
McAuliffe, Jon D. .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2017, 112 (518) :859-877
[8]   Latent Dirichlet allocation [J].
Blei, DM ;
Ng, AY ;
Jordan, MI .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (4-5) :993-1022
[9]  
Bouguila N, 2003, LECT NOTES ARTIF INT, V2734, P172
[10]   Unsupervised selection of a finite Dirichlet mixture model: An MML-based approach [J].
Bouguila, Nizar ;
Ziou, Djemel .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2006, 18 (08) :993-1009