Nonparametric Bayesian Deep Visualization

被引:0
|
作者
Ishizuka, Haruya [1 ]
Mochihashi, Daichi [2 ]
机构
[1] Bridgeston Corp, Chuo City, Japan
[2] Inst Stat Math, Tachikawa, Tokyo, Japan
来源
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT I | 2023年 / 13713卷
关键词
Data visualization; Gaussian processes; Nonparametric Bayesian models; Neural network; NONLINEAR DIMENSIONALITY REDUCTION; INFERENCE;
D O I
10.1007/978-3-031-26387-3_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visualization methods such as t-SNE [1] have helped in knowledge discovery from high-dimensional data; however, their performance may degrade when the intrinsic structure of observations is in low-dimensional space, and they cannot estimate clusters that are often useful to understand the internal structure of a dataset. A solution is to visualize the latent coordinates and clusters estimated using a neural clustering model. However, they require a long computational time since they have numerous weights to train and must tune the layer width, the number of latent dimensions and clusters to appropriately model the latent space. Additionally, the estimated coordinates may not be suitable for visualization since such a model and visualization method are applied independently. We utilize neural network Gaussian processes (NNGP) [2] equivalent to a neural network whose weights are marginalized to eliminate the necessity to optimize weights and layer widths. Additionally, to determine latent dimensions and the number of clusters without tuning, we propose a latent variable model that combines NNGP with automatic relevance determination [3] to extract necessary dimensions of latent space and infinite Gaussian mixture model [4] to infer the number of clusters. We integrate this model and visualization method into nonparametric Bayesian deep visualization (NPDV) that learns latent and visual coordinates jointly to render latent coordinates optimal for visualization. Experimental results on images and document datasets show that NPDV shows superior accuracy to existing methods, and it requires less training time than the neural clustering model because of its lower tuning cost. Furthermore, NPDV can reveal plausible latent clusters without labels.
引用
收藏
页码:121 / 137
页数:17
相关论文
共 50 条
  • [1] DNB: A Joint Learning Framework for Deep Bayesian Nonparametric Clustering
    Wang, Zeya
    Ni, Yang
    Jing, Baoyu
    Wang, Deqing
    Zhang, Hao
    Xing, Eric
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (12) : 7610 - 7620
  • [2] Bayesian regression with nonparametric heteroskedasticity
    Norets, Andriy
    JOURNAL OF ECONOMETRICS, 2015, 185 (02) : 409 - 419
  • [3] Discriminative Bayesian Nonparametric Clustering
    Nguyen, Vu
    Phung, Dinh
    Le, Trung
    Bui, Hung
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2550 - 2556
  • [4] Bayesian nonparametric hierarchical modeling
    Dunson, David B.
    BIOMETRICAL JOURNAL, 2009, 51 (02) : 273 - 284
  • [5] Bayesian nonparametric multiple testing
    Cipolli, William, III
    Hanson, Timothy
    McLain, Alexander C.
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2016, 101 : 64 - 79
  • [6] A Bayesian nonparametric causal model
    Karabatsos, George
    Walker, Stephen G.
    JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2012, 142 (04) : 925 - 934
  • [7] A tutorial on Bayesian nonparametric models
    Gershman, Samuel J.
    Blei, David M.
    JOURNAL OF MATHEMATICAL PSYCHOLOGY, 2012, 56 (01) : 1 - 12
  • [8] Bayesian nonparametric monotone regression
    Wilson, Ander
    Tryner, Jessica
    L'Orange, Christian
    Volckens, John
    ENVIRONMETRICS, 2020, 31 (08)
  • [9] Bayesian nonparametric estimation of a copula
    Wu, Juan
    Wang, Xue
    Walker, Stephen G.
    JOURNAL OF STATISTICAL COMPUTATION AND SIMULATION, 2015, 85 (01) : 103 - 116
  • [10] Approximate Bayesian Computation: A Nonparametric Perspective
    Blum, Michael G. B.
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2010, 105 (491) : 1178 - 1187