Graph Structure Aware Contrastive Multi-View Clustering

被引:6
作者
Chen, Rui [1 ,2 ]
Tang, Yongqiang [2 ]
Cai, Xiangrui [3 ]
Yuan, Xiaojie [4 ]
Feng, Wenlong [1 ,5 ]
Zhang, Wensheng [1 ,2 ]
机构
[1] Hainan Univ, Coll Informat Sci & Technol, Haikou 570228, Peoples R China
[2] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence S, Beijing 100190, Peoples R China
[3] Nankai Univ, Coll Comp Sci, Tianjin 300350, Peoples R China
[4] Nankai Univ, Coll Cyber Sci, Tianjin Key Lab Network & Data Secur Technol, Tianjin 300350, Peoples R China
[5] Hainan Univ, State Key Lab Marine Resource Utilizat South China, Haikou 570228, Peoples R China
关键词
Correlation; Semantics; Big Data; Representation learning; Data models; Data mining; Analytical models; Contrastive learning; deep representation; graph embedding; multi-view clustering; REPRESENTATION; ALGORITHM;
D O I
10.1109/TBDATA.2023.3334674
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-view clustering has become a research hotspot in recent decades because of its effectiveness in heterogeneous data fusion. Although a large number of related studies have been developed one after another, most of them usually only concern with the characteristics of the data themselves and overlook the inherent connection among samples, hindering them from exploring structural knowledge of graph space. Moreover, many current works tend to highlight the compactness of one cluster without taking the differences between clusters into account. To track these two drawbacks, in this article, we propose a graph structure aware contrastive multi-view clustering (namely, GCMC) approach. Specifically, we incorporate the well-designed graph autoencoder with conventional multi-layer perception autoencoder to extract the structural and high-level representation of multi-view data, so that the underlying correlation of samples can be effectively squeezed for model learning. Then the contrastive learning paradigm is performed on multiple pseudo-label distributions to ensure that the positive pairs of pseudo-label representations share the complementarity across views while the divergence between negative pairs is sufficiently large. This makes each semantic cluster more discriminative, i.e., jointly satisfying intra-cluster compactness and inter-cluster exclusiveness. Through comprehensive experiments on eight widely-known datasets, we prove that the proposed approach can perform better than the state-of-the-art opponents.
引用
收藏
页码:260 / 274
页数:15
相关论文
共 79 条
[1]  
Akaho S., 2007, arXiv
[2]  
Andrew G., 2013, INT C MACHINE LEARNI, V28, P1247
[3]   Multimodal Machine Learning: A Survey and Taxonomy [J].
Baltrusaitis, Tadas ;
Ahuja, Chaitanya ;
Morency, Louis-Philippe .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (02) :423-443
[4]  
Benton A, 2019, 4TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP (REPL4NLP-2019), P1
[5]  
Bishop C. M., 2006, MACH LEARN, DOI [DOI 10.18637/JSS.V017.B05, 10.5555/1162264]
[6]   Structural Deep Clustering Network [J].
Bo, Deyu ;
Wang, Xiao ;
Shi, Chuan ;
Zhu, Meiqi ;
Lu, Emiao ;
Cui, Peng .
WEB CONFERENCE 2020: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2020), 2020, :1400-1410
[7]  
Cai X., 2013, P 23 INT JOINT C ART, DOI [10.5555/2540128.2540503, DOI 10.5555/2540128.2540503]
[8]   Incomplete multi-view clustering with multiple imputation and ensemble clustering [J].
Chao, Guoqing ;
Wang, Songtao ;
Yang, Shiming ;
Li, Chunshan ;
Chu, Dianhui .
APPLIED INTELLIGENCE, 2022, 52 (13) :14811-14821
[9]  
Chao Guoqing, 2021, IEEE Trans Artif Intell, V2, P146, DOI 10.1109/tai.2021.3065894
[10]   Semisupervised Progressive Representation Learning for Deep Multiview Clustering [J].
Chen, Rui ;
Tang, Yongqiang ;
Xie, Yuan ;
Feng, Wenlong ;
Zhang, Wensheng .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (10) :14341-14355