Bayesian mixture models (in)consistency for the number of clusters

被引:1
|
作者
Alamichel, Louise [1 ]
Bystrova, Daria [1 ,2 ]
Arbel, Julyan [1 ]
King, Guillaume Kon Kam [3 ]
机构
[1] Univ Grenoble Alpes, Inria, Grenoble INP, LJK,CNRS, Grenoble, France
[2] Univ Savoie Mont Blanc, CNRS, Lab Ecol Alpine, Univ Grenoble Alpes, Grenoble, France
[3] Univ Paris Saclay, INRAE, MaIAGE, Jouy En Josas, France
关键词
clustering; finite mixtures; finite-dimensional BNP representations; Gibbs-type process; GIBBS-TYPE PRIORS; PITMAN-YOR; NONPARAMETRIC-INFERENCE; DIRICHLET MIXTURES; DENSITY-ESTIMATION; CONVERGENCE-RATES; FINITE; CONSISTENCY;
D O I
10.1111/sjos.12739
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Bayesian nonparametric mixture models are common for modeling complex data. While these models are well-suited for density estimation, recent results proved posterior inconsistency of the number of clusters when the true number of components is finite, for the Dirichlet process and Pitman-Yor process mixture models. We extend these results to additional Bayesian nonparametric priors such as Gibbs-type processes and finite-dimensional representations thereof. The latter include the Dirichlet multinomial process, the recently proposed Pitman-Yor, and normalized generalized gamma multinomial processes. We show that mixture models based on these processes are also inconsistent in the number of clusters and discuss possible solutions. Notably, we show that a postprocessing algorithm introduced for the Dirichlet process can be extended to more general models and provides a consistent method to estimate the number of components.
引用
收藏
页码:1619 / 1660
页数:42
相关论文
共 50 条
  • [41] On a loss-based prior for the number of components in mixture models
    Grazian, Clara
    Villa, Cristiano
    Liseo, Brunero
    STATISTICS & PROBABILITY LETTERS, 2020, 158
  • [42] Bayesian approaches to variable selection in mixture models with application to disease clustering
    Lu, Zihang
    Lou, Wendy
    JOURNAL OF APPLIED STATISTICS, 2023, 50 (02) : 387 - 407
  • [43] On sufficient conditions for Bayesian consistency
    Walker, S
    BIOMETRIKA, 2003, 90 (02) : 482 - 488
  • [44] ADAPTIVE NONPARAMETRIC BAYESIAN INFERENCE USING LOCATION-SCALE MIXTURE PRIORS
    de Jonge, R.
    van Zanten, J. H.
    ANNALS OF STATISTICS, 2010, 38 (06) : 3300 - 3320
  • [45] Automatic selection of the number of clusters using Bayesian clustering and sparsity-inducing priors
    Valle, Denis
    Jameel, Yusuf
    Betancourt, Brenda
    Azeria, Ermias T.
    Attias, Nina
    Cullen, Joshua
    ECOLOGICAL APPLICATIONS, 2022, 32 (03)
  • [46] Consistency of discrete Bayesian learning
    Poland, Jan
    THEORETICAL COMPUTER SCIENCE, 2008, 405 (03) : 256 - 273
  • [47] A Bayesian mixture of experts approach to covariate misclassification
    Xia, Michelle
    Hahn, P. Richard
    Gustafson, Paul
    CANADIAN JOURNAL OF STATISTICS-REVUE CANADIENNE DE STATISTIQUE, 2020, 48 (04): : 731 - 750
  • [48] POSTERIOR GRAPH SELECTION AND ESTIMATION CONSISTENCY FOR HIGH-DIMENSIONAL BAYESIAN DAG MODELS
    Cao, Xuan
    Khare, Kshitij
    Ghosh, Malay
    ANNALS OF STATISTICS, 2019, 47 (01) : 319 - 348
  • [49] Bayesian estimation of the number of change points in simple linear regression models
    Fan, TH
    Chang, KC
    Lee, CB
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2006, 35 (04) : 689 - 710
  • [50] On Bayesian mixture credibility
    Lau, John W.
    Siu, Tak Kuen
    Yang, Hailiang
    ASTIN BULLETIN, 2006, 36 (02): : 573 - 588