Multi-task self-supervised learning based fusion representation for Multi-view clustering

被引:0
作者
Guo, Tianlong [1 ]
Shen, Derong [1 ]
Kou, Yue [1 ]
Nie, Tiezheng [1 ]
机构
[1] Northeastern Univ, Sch Comp Sci & Engn, Shenyang, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-view; Multi-task; Clustering; Representation learning; Unsupervised learning; Information fusion;
D O I
10.1016/j.ins.2024.121705
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-view data clustering aims to utilize the complementary information from different views and is widely studied in unsupervised learning. However, it is a significant challenge to extract the common and view-specific representations within a single framework. Concurrently, Multi-task Learning (MTL) excels at learning common and specific representations from multiple inputs, but it is primarily designed for supervised learning. Direct applying of existing MTL training methods to unsupervised scenarios is not feasible due to the lack of labels. To address these challenges, we propose a Multi-task Self-supervised Learning based Multi-view Fusion Representation (MSLMFR) model, which adapts the MTL frameworks to unsupervised scenarios, with three types of tasks specifically designed. Distinct from conventional MTL frameworks, our main tasks focus on the fusion representation and reconstruction of different views. Additionally, an auxiliary anchor task is introduced to further enhance performance. Subsequently, we propose a multi-task gradient optimization method. This method updates the gradient through matrix alignment in both parameter and task spaces, and takes a norm for each task. We conduct experiments on six real- world datasets, including both images and texts. The results demonstrate the effectiveness and efficiency of MSLMFR in learning high-quality fusion representations.
引用
收藏
页数:10
相关论文
共 37 条
[11]  
Kang Z, 2020, AAAI CONF ARTIF INTE, V34, P4412
[12]  
Kumar A., 2011, P 28 INT C MACH LEAR, P393, DOI DOI 10.5555/3104482.3104532
[13]  
Kumar A., 2011, P ADV NEUR INF PROC, V24, DOI [10.5555/2986459.2986617, DOI 10.5555/2986459.2986617]
[14]   Multi-view representation model based on graph autoencoder [J].
Li, Jingci ;
Lu, Guangquan ;
Wu, Zhengtian ;
Ling, Fuqing .
INFORMATION SCIENCES, 2023, 632 :439-453
[15]  
Lin BJ, 2022, Arxiv, DOI [arXiv:2111.10603, 10.48550/ARXIV.2111.10603, DOI 10.48550/ARXIV.2111.10603]
[16]  
Liu Bo, 2021, Advances in Neural Information Processing Systems, V34
[17]  
Liu Lu, 2021, iclr
[18]   End-to-End Multi-Task Learning with Attention [J].
Liu, Shikun ;
Johns, Edward ;
Davison, Andrew J. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :1871-1880
[19]   Bidirectional Attentive Multi-View Clustering [J].
Lu, Jitao ;
Nie, Feiping ;
Dong, Xia ;
Wang, Rong ;
Li, Xuelong .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (05) :1889-1901
[20]   Multi-view subspace clustering via partition fusion [J].
Lv, Juncheng ;
Kang, Zhao ;
Wang, Boyu ;
Ji, Luping ;
Xu, Zenglin .
INFORMATION SCIENCES, 2021, 560 (560) :410-423