Short text clustering based on Pitman-Yor process mixture model

被引:29
|
作者
Qiang, Jipeng [1 ]
Li, Yun [1 ]
Yuan, Yunhao [1 ]
Wu, Xindong [2 ,3 ]
机构
[1] Yangzhou Univ, Dept Comp Sci, Yangzhou, Jiangsu, Peoples R China
[2] Hefei Univ Technol, Dept Comp Sci, Hefei, Anhui, Peoples R China
[3] Univ Louisiana Lafayette, Sch Comp & Informat, Lafayette, LA 70504 USA
基金
中国国家自然科学基金;
关键词
LDA; Pitman-Yor process; Short text clustering; NONNEGATIVE MATRIX FACTORIZATION; ALGORITHMS;
D O I
10.1007/s10489-017-1055-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For finding the appropriate number of clusters in short text clustering, models based on Dirichlet Multinomial Mixture (DMM) require the maximum possible cluster number before inferring the real number of clusters. However, it is difficult to choose a proper number as we do not know the true number of clusters in short texts beforehand. The cluster distribution in DMM based on Dirichlet process as prior goes down exponentially as the number of clusters increases. Therefore, we propose a novel model based on Pitman-Yor Process to capture the power-law phenomenon of the cluster distribution in the paper. Specifically, each text chooses one of the active clusters or a new cluster with probabilities derived from the Pitman-Yor Process Mixture model (PYPM). Discriminative words and nondiscriminative words are identified automatically to help enhance text clustering. Parameters are estimated efficiently by collapsed Gibbs sampling and experimental results show PYPM is robust and effective comparing with the state-of-the-art models.
引用
收藏
页码:1802 / 1812
页数:11
相关论文
共 50 条
  • [21] A latent variable Gaussian process model with Pitman-Yor process priors for multiclass classification
    Chatzis, Sotirios P.
    NEUROCOMPUTING, 2013, 120 : 482 - 489
  • [22] Parallel Markov Chain Monte Carlo for Pitman-Yor Mixture Models
    Dubey, Avinava
    Williamson, Sinead A.
    Xing, Eric P.
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2014, : 142 - 151
  • [23] Spatial emission tomography reconstruction using Pitman-Yor process
    Fall, Mame Diarra
    Barat, Eric
    Mohammad-Djafari, Ali
    Comtat, Claude
    BAYESIAN INFERENCE AND MAXIMUM ENTROPY METHODS IN SCIENCE AND ENGINEERING, 2009, 1193 : 194 - +
  • [24] Online Learning of Hierarchical Pitman-Yor Process Mixture of Generalized Dirichlet Distributions With Feature Selection
    Fan, Wentao
    Sallay, Hassen
    Bouguila, Nizar
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (09) : 2048 - 2061
  • [25] BAYESIAN COMMON SPATIAL PATTERNS WITH PITMAN-YOR PROCESS PRIORS
    Kang, Hyohyeong
    Choi, Seungjin
    2012 PROCEEDINGS OF THE 20TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2012, : 684 - 688
  • [26] Hidden Markov model with Pitman-Yor priors for probabilistic topic model
    Guo, Jianjie
    Guo, Lin
    Xu, Wenchao
    Zhang, Haibin
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2025, 54 (09) : 2791 - 2805
  • [27] A Markov random field-regulated Pitman-Yor process prior for spatially constrained data clustering
    Chatzis, Sotirios P.
    PATTERN RECOGNITION, 2013, 46 (06) : 1595 - 1603
  • [28] A Parallel Training Algorithm for Hierarchical Pitman-Yor Process Language Models
    Huang, Songfang
    Renals, Steve
    INTERSPEECH 2009: 10TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2009, VOLS 1-5, 2009, : 2663 - 2666
  • [29] Bernstein-von Mises theorem for the Pitman-Yor process of nonnegative
    Franssen, S. E. M. P.
    van der Vaart, A. W.
    ELECTRONIC JOURNAL OF STATISTICS, 2022, 16 (02): : 5779 - 5811
  • [30] Simultaneous clustering and feature selection via nonparametric Pitman–Yor process mixture models
    Wentao Fan
    Nizar Bouguila
    International Journal of Machine Learning and Cybernetics, 2019, 10 : 2753 - 2766