Multi-VAE: Learning Disentangled View-common and View-peculiar Visual Representations for Multi-view Clustering

被引:75
|
作者
Xu, Jie [1 ]
Ren, Yazhou [1 ]
Tang, Huayi [1 ]
Pu, Xiaorong [1 ]
Zhu, Xiaofeng [1 ]
Zeng, Ming [2 ]
He, Lifang [3 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Peoples R China
[2] Carnegie Mellon Univ, Dept Elect Comp Engn, Pittsburgh, PA 15213 USA
[3] Lehigh Univ, Dept Comp Sci & Engn, Bethlehem, PA 18015 USA
来源
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021) | 2021年
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV48922.2021.00910
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-view clustering, a long-standing and important research problem, focuses on mining complementary information from diverse views. However, existing works often fuse multiple views' representations or handle clustering in a common feature space, which may result in their entanglement especially for visual representations. To address this issue, we present a novel VAE-based multi-view clustering framework (Multi-VAE) by learning disentangled visual representations. Concretely, we define a view-common variable and multiple view-peculiar variables in the generative model. The prior of view-common variable obeys approximately discrete Gumbel Softmax distribution, which is introduced to extract the common cluster factor of multiple views. Meanwhile, the prior of view-peculiar variable follows continuous Gaussian distribution, which is used to represent each view's peculiar visual factors. By controlling the mutual information capacity to disentangle the view-common and view-peculiar representations, continuous visual information of multiple views can be separated so that their common discrete cluster information can be effectively mined. Experimental results demonstrate that Multi-VAE enjoys the disentangled and explainable visual representations, while obtaining superior clustering performance compared with state-of-the-art methods.
引用
收藏
页码:9214 / 9223
页数:10
相关论文
共 50 条
  • [1] Divergence-guided disentanglement of view-common and view-unique representations for multi-view data
    Lu, Mingfei
    Zhang, Qi
    Chen, Badong
    INFORMATION FUSION, 2025, 114
  • [2] Learning latent disentangled embeddings and graphs for multi-view clustering
    Zhang, Chao
    Chen, Haoxing
    Li, Huaxiong
    Chen, Chunlin
    PATTERN RECOGNITION, 2024, 156
  • [3] Fast Disentangled Slim Tensor Learning for Multi-View Clustering
    Xu, Deng
    Zhang, Chao
    Li, Zechao
    Chen, Chunlin
    Li, Huaxiong
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 1254 - 1265
  • [4] Multi-view Proximity Learning for Clustering
    Lin, Kun-Yu
    Huang, Ling
    Wang, Chang-Dong
    Chao, Hong-Yang
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2018), PT II, 2018, 10828 : 407 - 423
  • [5] Multi-view clustering
    Bickel, S
    Scheffer, T
    FOURTH IEEE INTERNATIONAL CONFERENCE ON DATA MINING, PROCEEDINGS, 2004, : 19 - 26
  • [6] Multi-view clustering based on graph learning and view diversity learning
    Lin Wang
    Dong Sun
    Zhu Yuan
    Qingwei Gao
    Yixiang Lu
    The Visual Computer, 2023, 39 : 6133 - 6149
  • [7] Learning consensus representations in multi-latent spaces for multi-view clustering
    Ma, Qianli
    Zheng, Jiawei
    Li, Sen
    Zheng, Zhenjing
    Cottrell, Garrison W.
    NEUROCOMPUTING, 2024, 596
  • [8] Multi-view clustering based on graph learning and view diversity learning
    Wang, Lin
    Sun, Dong
    Yuan, Zhu
    Gao, Qingwei
    Lu, Yixiang
    VISUAL COMPUTER, 2023, 39 (12): : 6133 - 6149
  • [9] Diverse and Common Multi-View Subspace Clustering
    Lu, Zhiqiang
    Wu, Songsong
    Liu, Yurong
    Gao, Guangwei
    Wu, Fei
    PROCEEDINGS OF 2018 5TH IEEE INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND INTELLIGENCE SYSTEMS (CCIS), 2018, : 878 - 882
  • [10] Multi-view representation learning for multi-view action recognition
    Hao, Tong
    Wu, Dan
    Wang, Qian
    Sun, Jin-Sheng
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2017, 48 : 453 - 460