Deep multi-view semi-supervised clustering with sample pairwise constraints

被引:23
作者
Chen, Rui [1 ,2 ,3 ]
Tang, Yongqiang [2 ]
Zhang, Wensheng [1 ,2 ]
Feng, Wenlong [1 ,3 ]
机构
[1] Hainan Univ, Coll Informat Sci & Technol, Haikou, Peoples R China
[2] Chinese Acad Sci, Inst Automat, Res Ctr Precis Sensing & Control, Beijing, Peoples R China
[3] Hainan Univ, State Key Lab Marine Resource Utilizat South China, Haikou, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-view clustering; Deep clustering; Semi-supervised clustering; Pairwise constraints; Feature properties protection; SCALE;
D O I
10.1016/j.neucom.2022.05.091
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-view clustering has attracted much attention thanks to the capacity of multi-source information integration. Although numerous advanced methods have been proposed in past decades, most of them generally overlook the significance of weakly-supervised information and fail to preserve the feature properties of multiple views, thus resulting in unsatisfactory clustering performance. To address these issues, in this paper, we propose a novel Deep Multi-view Semi-supervised Clustering (DMSC) method, which jointly optimizes three kinds of losses during networks finetuning, including multi-view clustering loss, semi-supervised pairwise constraint loss and multiple autoencoders reconstruction loss. Specifically, a KL divergence based multi-view clustering loss is imposed on the common representation of multiview data to perform heterogeneous feature optimization, multi-view weighting and clustering prediction simultaneously. Then, we innovatively propose to integrate pairwise constraints into the process of multi-view clustering by enforcing the learned multi-view representation of must-link samples (cannot-link samples) to be similar (dissimilar), such that the formed clustering architecture can be more credible. Moreover, unlike existing rivals that only preserve the encoders for each heterogeneous branch during networks finetuning, we further propose to tune the intact autoencoders frame that contains both encoders and decoders. In this way, the issue of serious corruption of view-specific and view-shared feature space could be alleviated, making the whole training procedure more stable. Through comprehensive experiments on eight popular image datasets, we demonstrate that our proposed approach performs better than the state-of-the-art multi-view and single-view competitors. (c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:832 / 845
页数:14
相关论文
共 62 条
[41]   Semi-supervised deep embedded clustering [J].
Ren, Yazhou ;
Hu, Kangrong ;
Dai, Xinyi ;
Pan, Lili ;
Hoi, Steven C. H. ;
Xu, Zenglin .
NEUROCOMPUTING, 2019, 325 :121-130
[42]   Normalized cuts and image segmentation [J].
Shi, JB ;
Malik, J .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2000, 22 (08) :888-905
[43]   Face Clustering: Representation and Pairwise Constraints [J].
Shi, Yichun ;
Otto, Charles ;
Jain, Anil K. .
IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2018, 13 (07) :1626-1640
[44]  
Strehl A, 2002, EIGHTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-02)/FOURTEENTH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE (IAAI-02), PROCEEDINGS, P93, DOI 10.1162/153244303321897735
[45]  
Tang Y, 2021, IEEE T MULTIMEDIA
[46]   One-Step Multiview Subspace Segmentation via Joint Skinny Tensor Learning and Latent Clustering [J].
Tang, Yongqiang ;
Xie, Yuan ;
Zhang, Changqing ;
Zhang, Zhizhong ;
Zhang, Wensheng .
IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (09) :9179-9193
[47]   Tensor Multi-Elastic Kernel Self-Paced Learning for Time Series Clustering [J].
Tang, Yongqiang ;
Xie, Yuan ;
Yang, Xuebing ;
Niu, Jinghao ;
Zhang, Wensheng .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2021, 33 (03) :1223-1237
[48]  
van der Maaten L, 2008, J MACH LEARN RES, V9, P2579
[49]  
Wang WR, 2015, PR MACH LEARN RES, V37, P1083
[50]  
Wen J, 2020, PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P3230