Bayesian mixture models (in)consistency for the number of clusters

被引:1
|
作者
Alamichel, Louise [1 ]
Bystrova, Daria [1 ,2 ]
Arbel, Julyan [1 ]
King, Guillaume Kon Kam [3 ]
机构
[1] Univ Grenoble Alpes, Inria, Grenoble INP, LJK,CNRS, Grenoble, France
[2] Univ Savoie Mont Blanc, CNRS, Lab Ecol Alpine, Univ Grenoble Alpes, Grenoble, France
[3] Univ Paris Saclay, INRAE, MaIAGE, Jouy En Josas, France
关键词
clustering; finite mixtures; finite-dimensional BNP representations; Gibbs-type process; GIBBS-TYPE PRIORS; PITMAN-YOR; NONPARAMETRIC-INFERENCE; DIRICHLET MIXTURES; DENSITY-ESTIMATION; CONVERGENCE-RATES; FINITE; CONSISTENCY;
D O I
10.1111/sjos.12739
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Bayesian nonparametric mixture models are common for modeling complex data. While these models are well-suited for density estimation, recent results proved posterior inconsistency of the number of clusters when the true number of components is finite, for the Dirichlet process and Pitman-Yor process mixture models. We extend these results to additional Bayesian nonparametric priors such as Gibbs-type processes and finite-dimensional representations thereof. The latter include the Dirichlet multinomial process, the recently proposed Pitman-Yor, and normalized generalized gamma multinomial processes. We show that mixture models based on these processes are also inconsistent in the number of clusters and discuss possible solutions. Notably, we show that a postprocessing algorithm introduced for the Dirichlet process can be extended to more general models and provides a consistent method to estimate the number of components.
引用
收藏
页码:1619 / 1660
页数:42
相关论文
共 50 条
  • [21] Optimal Bayesian estimation of Gaussian mixtures with growing number of components
    Ohn, Ilsang
    Lin, Lizhen
    BERNOULLI, 2023, 29 (02) : 1195 - 1218
  • [22] BAYESIAN CONSISTENCY WITH THE SUPREMUM METRIC
    Ho, Nhat
    Walker, Stephen G.
    STATISTICA SINICA, 2024, 34 (03) : 1281 - 1297
  • [23] A novel approach to Bayesian consistency
    Chae, Minwoo
    Walker, Stephen G.
    ELECTRONIC JOURNAL OF STATISTICS, 2017, 11 (02): : 4723 - 4745
  • [24] FELLER OPERATORS AND MIXTURE PRIORS IN BAYESIAN NONPARAMETRICS
    Petrone, Sonia
    Veronese, Piero
    STATISTICA SINICA, 2010, 20 (01) : 379 - 404
  • [25] Infinite Gaussian Mixture Modeling with an Improved Estimation of the Number of Clusters
    Matza, Avi
    Bistritz, Yuval
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8921 - 8929
  • [26] GPGPU Implementation of Variational Bayesian Gaussian Mixture Models
    Nishimoto, Hiroki
    Zhang, Renyuan
    Nakashima, Yasuhiko
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (03) : 611 - 622
  • [27] The Sensitivity of the Number of Clusters in a Gaussian Mixture Model to Prior Distributions
    Cruz, Cristian
    Leao, William Lima
    Rohde, David
    MATHEMATICS IN COMPUTER SCIENCE, 2013, 7 (04) : 401 - 420
  • [28] Bayesian Dependent Mixture Models: A Predictive Comparison and Survey
    Wade, Sara
    Inacio, Vanda
    STATISTICAL SCIENCE, 2025, 40 (01) : 81 - 108
  • [29] GPGPU Implementation of Variational Bayesian Gaussian Mixture Models
    Nishimoto, Hiroki
    Nakada, Takashi
    Nakashima, Yasuhiko
    2019 SEVENTH INTERNATIONAL SYMPOSIUM ON COMPUTING AND NETWORKING (CANDAR 2019), 2019, : 185 - 190
  • [30] Fully Bayesian Learning of Multivariate Beta Mixture Models
    Amirkhani, Mahsa
    Manouchehri, Narges
    Bouguila, Nizar
    2020 IEEE 21ST INTERNATIONAL CONFERENCE ON INFORMATION REUSE AND INTEGRATION FOR DATA SCIENCE (IRI 2020), 2020, : 120 - 127