Partially View-Aligned Representation Learning via Cross-View Graph Contrastive Network

被引:4
作者
Wang, Yiming [1 ]
Chang, Dongxia [2 ,3 ]
Fu, Zhiqiang [4 ]
Wen, Jie [5 ]
Zhao, Yao [2 ]
机构
[1] Nanjing Univ Posts & Telecommun, Sch Comp Sci, Nanjing 210023, Peoples R China
[2] Beijing Jiaotong Univ, Inst Informat Sci, Beijing 100044, Peoples R China
[3] Network Technol, Beijing Key Lab Adv Informat Sci, Beijing 100044, Peoples R China
[4] China Construct Bank, Beijing 100033, Peoples R China
[5] Harbin Inst Technol, Shenzhen Key Lab Visual Object Detect & Recognit, Shenzhen 150001, Peoples R China
基金
中国国家自然科学基金;
关键词
Representation learning; Self-supervised learning; Task analysis; Correlation; Circuits and systems; Measurement; Visualization; Multi-view representation learning; partial view-aligned multi-view learning; contrastive learning;
D O I
10.1109/TCSVT.2024.3376720
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Multi-view representation learning, aimed at uncovering the inherent structure within multi-view data, has developed rapidly in recent years. In practice, due to temporal and spatial desynchronization, it is common that only part of the data is aligned between views, which leads to the Partial View Alignment (PVA) problem. To address the challenge of representation learning on partially view-aligned multi-view data, we propose a new cross-view graph contrastive learning network, which integrates multi-view information to align data and learn latent representations. First, view-specific autoencoders are used to construct an end-to-end multi-view representation learning framework for learning specific view representations. Furthermore, to achieve cluster-level alignment, we introduce a cross-view graph contrastive learning module to guide the learning of discriminative representations. Compared to the existing methods, the proposed cluster-level alignment method successfully extends the view alignment to more than two views. Meanwhile, the results of clustering and classification experiments on several popular multi-view datasets can also illustrate the effectiveness and superiority of the proposed method.
引用
收藏
页码:7272 / 7283
页数:12
相关论文
共 56 条
[1]  
Andrew G., 2013, INT C MACHINE LEARNI, V28, P1247
[2]  
[Anonymous], 2009, Adv. Neural Inf. Process. Syst.
[3]  
Arthur D, 2007, PROCEEDINGS OF THE EIGHTEENTH ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS, P1027
[4]  
Chen T, 2020, PR MACH LEARN RES, V119
[5]   Low-Rank Tensor Graph Learning for Multi-View Subspace Clustering [J].
Chen, Yongyong ;
Xiao, Xiaolin ;
Peng, Chong ;
Lu, Guangming ;
Zhou, Yicong .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (01) :92-104
[6]  
d'Aspremont A., 2007, P 24 INT C MACHINE L, P177
[7]  
Fei-Fei L, 2005, PROC CVPR IEEE, P524
[8]   Efficient and Distributed Generalized Canonical Correlation Analysis for Big Multiview Data [J].
Fu, Xiao ;
Huang, Kejun ;
Papalexakis, Evangelos E. ;
Song, Hyun Ah ;
Talukdar, Partha ;
Sidiropoulos, Nicholas D. ;
Faloutsos, Christos ;
Mitchell, Tom .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2019, 31 (12) :2304-2318
[9]   Pairwise Two-Stream ConvNets for Cross-Domain Action Recognition With Small Data [J].
Gao, Zan ;
Guo, Leming ;
Ren, Tongwei ;
Liu, An-An ;
Cheng, Zhi-Yong ;
Chen, Shengyong .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (03) :1147-1161
[10]  
Goodfellow I, 2014, arXiv