共 50 条
AMCFCN: attentive multi-view contrastive fusion clustering net
被引:0
|作者:
Xiao, Huarun
[1
]
Hong, Zhiyong
[1
]
Xiong, Liping
[1
]
Zeng, Zhiqiang
[1
]
机构:
[1] Wuyi Univ, Coll Elect & Informat Engn, Jiangmen, Guangdong, Peoples R China
基金:
中国国家自然科学基金;
关键词:
Multi-view Clustering;
Contrastive Learning;
Attention Mechanism;
D O I:
10.7717/peerj-cs.1906
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
Advances in deep learning have propelled the evolution of multi-view clustering techniques, which strive to obtain a view-common representation from multi-view datasets. However, the contemporary multi-view clustering community confronts two prominent challenges. One is that view-specific representations lack guarantees to reduce noise introduction, and another is that the fusion process compromises view-specific representations, resulting in the inability to capture efficient information from multi-view data. This may negatively affect the accuracy of the clustering results. In this article, we introduce a novel technique named the "contrastive attentive strategy"to address the above problems. Our approach effectively extracts robust viewspecific representations from multi-view data with reduced noise while preserving view completeness. This results in the extraction of consistent representations from multi-view data while preserving the features of view-specific representations. We integrate view-specific encoders, a hybrid attentive module, a fusion module, and deep clustering into a unified framework called AMCFCN. Experimental results on four multi-view datasets demonstrate that our method, AMCFCN, outperforms seven competitive multi-view clustering methods. Our source code is available at https: //github.com/xiaohuarun/AMCFCN.
引用
收藏
页码:1 / 25
页数:25
相关论文