A kernel Stein test for comparing latent variable models

被引:1
作者
Kanagawa, Heishiro [1 ]
Jitkrittum, Wittawat [2 ,3 ]
Mackey, Lester
Fukumizu, Kenji [4 ]
Gretton, Arthur [5 ]
机构
[1] UCL, Gatsby Computat Neurosci Unit, London, England
[2] Max Planck Inst Intelligent Syst, Empir Inference, Tubingen, Germany
[3] Google Res, New York, NY USA
[4] Microsoft Res New England, Cambridge, MA USA
[5] Inst Stat Math, Tachikawa, Tokyo, Japan
关键词
hypothesis testing; kernel methods; mixture models; model selection; Stein's method; STATIONARY DISTRIBUTIONS; GAMMA-DISTRIBUTION; CONVERGENCE;
D O I
10.1093/jrsssb/qkad050
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We propose a kernel-based nonparametric test of relative goodness of fit, where the goal is to compare two models, both of which may have unobserved latent variables, such that the marginal distribution of the observed variables is intractable. The proposed test generalizes the recently proposed kernel Stein discrepancy (KSD) tests (Liu et al., Proceedings of the 33rd international conference on machine learning (pp. 276-284); Chwialkowski et al., (2016), In Proceedings of the 33rd international conference on machine learning (pp. 2606-2615); Yang et al., (2018), In Proceedings of the 35th international conference on machine learning (pp. 5561-5570)) to the case of latent variable models, a much more general class than the fully observed models treated previously. The new test, with a properly calibrated threshold, has a well-controlled type-I error. In the case of certain models with low-dimensional latent structures and high-dimensional observations, our test significantly outperforms the relative maximum mean discrepancy test, which is based on samples from the models and does not exploit the latent structure.
引用
收藏
页码:986 / 1011
页数:26
相关论文
共 84 条
  • [11] Testing normality via a distributional fixed point property in the Stein characterization
    Betsch, Steffen
    Ebner, Bruno
    [J]. TEST, 2020, 29 (01) : 105 - 138
  • [12] A new characterization of the Gamma distribution and associated goodness-of-fit tests
    Betsch, Steffen
    Ebner, Bruno
    [J]. METRIKA, 2019, 82 (07) : 779 - 806
  • [13] Latent Dirichlet allocation
    Blei, DM
    Ng, AY
    Jordan, MI
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (4-5) : 993 - 1022
  • [14] Graph Kernels
    Borgwardt, Karsten
    Ghisu, Elisabetta
    Llinares-Lopez, Felipe
    O'Bray, Leslie
    Rieck, Bastian
    [J]. FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2020, 13 (5-6): : 531 - 712
  • [15] COUPLING AND CONVERGENCE FOR HAMILTONIAN MONTE CARLO
    Bou-Rabee, Nawaf
    Eberle, Andreas
    Zimmer, Raphael
    [J]. ANNALS OF APPLIED PROBABILITY, 2020, 30 (03) : 1209 - 1250
  • [16] Bounliphone Wacha, 2016, 4 INT C LEARNING REP
  • [17] STEIN'S METHOD FOR STATIONARY DISTRIBUTIONS OF MARKOV CHAINS AND APPLICATION TO ISING MODELS
    Bresler, Guy
    Nagaraj, Dheeraj
    [J]. ANNALS OF APPLIED PROBABILITY, 2019, 29 (05) : 3230 - 3265
  • [18] POISSON APPROXIMATION FOR DEPENDENT TRIALS
    CHEN, LHY
    [J]. ANNALS OF PROBABILITY, 1975, 3 (03) : 534 - 545
  • [19] Chen WY, 2019, PR MACH LEARN RES, V97
  • [20] Christmann A., 2008, Support Vector Machines