Multi-VAE: Learning Disentangled View-common and View-peculiar Visual Representations for Multi-view Clustering

被引:75
|
作者
Xu, Jie [1 ]
Ren, Yazhou [1 ]
Tang, Huayi [1 ]
Pu, Xiaorong [1 ]
Zhu, Xiaofeng [1 ]
Zeng, Ming [2 ]
He, Lifang [3 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Peoples R China
[2] Carnegie Mellon Univ, Dept Elect Comp Engn, Pittsburgh, PA 15213 USA
[3] Lehigh Univ, Dept Comp Sci & Engn, Bethlehem, PA 18015 USA
来源
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021) | 2021年
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV48922.2021.00910
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-view clustering, a long-standing and important research problem, focuses on mining complementary information from diverse views. However, existing works often fuse multiple views' representations or handle clustering in a common feature space, which may result in their entanglement especially for visual representations. To address this issue, we present a novel VAE-based multi-view clustering framework (Multi-VAE) by learning disentangled visual representations. Concretely, we define a view-common variable and multiple view-peculiar variables in the generative model. The prior of view-common variable obeys approximately discrete Gumbel Softmax distribution, which is introduced to extract the common cluster factor of multiple views. Meanwhile, the prior of view-peculiar variable follows continuous Gaussian distribution, which is used to represent each view's peculiar visual factors. By controlling the mutual information capacity to disentangle the view-common and view-peculiar representations, continuous visual information of multiple views can be separated so that their common discrete cluster information can be effectively mined. Experimental results demonstrate that Multi-VAE enjoys the disentangled and explainable visual representations, while obtaining superior clustering performance compared with state-of-the-art methods.
引用
收藏
页码:9214 / 9223
页数:10
相关论文
共 50 条
  • [41] Multi-View Perceptron: a Deep Model for Learning Face Identity and View Representations
    Zhu, Zhenyao
    Luo, Ping
    Wang, Xiaogang
    Tang, Xiaoou
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [42] Multi-view dreaming: multi-view world model with contrastive learning
    Kinose A.
    Okumura R.
    Okada M.
    Taniguchi T.
    Advanced Robotics, 2023, 37 (19) : 1212 - 1220
  • [43] Multi-View Guided Multi-View Stereo
    Poggi, Matteo
    Conti, Andrea
    Mattoccia, Stefano
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 8391 - 8398
  • [44] Projected cross-view learning for unbalanced incomplete multi-view clustering
    Cai, Yiran
    Che, Hangjun
    Pan, Baicheng
    Leung, Man-Fai
    Liu, Cheng
    Wen, Shiping
    INFORMATION FUSION, 2024, 105
  • [45] Multi-Task Multi-View Clustering
    Zhang, Xiaotong
    Zhang, Xianchao
    Liu, Han
    Liu, Xinyue
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2016, 28 (12) : 3324 - 3338
  • [46] Multi-view Subspace Clustering via Joint Latent Representations
    Dong, Wenhua
    Wu, Xiao-jun
    Xu, Tianyang
    NEURAL PROCESSING LETTERS, 2022, 54 (03) : 1879 - 1901
  • [47] Multi-view Subspace Clustering via Joint Latent Representations
    Wenhua Dong
    Xiao-jun Wu
    Tianyang Xu
    Neural Processing Letters, 2022, 54 : 1879 - 1901
  • [48] Multi-view spectral clustering via common structure maximization of local and global representations
    Hao, Wenyu
    Pang, Shanmin
    Chen, Zhikai
    NEURAL NETWORKS, 2021, 143 : 595 - 606
  • [49] Learning enhanced specific representations for multi-view feature learning
    Hao, Yaru
    Jing, Xiao-Yuan
    Chen, Runhang
    Liu, Wei
    KNOWLEDGE-BASED SYSTEMS, 2023, 272
  • [50] Multi-view clustering via adversarial view embedding and adaptive view fusion
    Li, Yongzhen
    Liao, Husheng
    APPLIED INTELLIGENCE, 2021, 51 (03) : 1201 - 1212