Semisupervised Progressive Representation Learning for Deep Multiview Clustering

被引:10
作者
Chen, Rui [1 ,2 ]
Tang, Yongqiang [2 ]
Xie, Yuan [3 ]
Feng, Wenlong [1 ,4 ]
Zhang, Wensheng [1 ,2 ]
机构
[1] Hainan Univ, Coll Informat Sci & Technol, Haikou 570208, Peoples R China
[2] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence S, Beijing 100190, Peoples R China
[3] East China Normal Univ, Sch Comp Sci & Technol, Shanghai 200241, Peoples R China
[4] Hainan Univ, State Key Lab Marine Resource Utilizat South China, Haikou 570208, Peoples R China
基金
上海市自然科学基金; 中国国家自然科学基金;
关键词
Representation learning; Training; Data models; Task analysis; Complexity theory; Semisupervised learning; Optimization; Deep clustering; multiview clustering; progressive sample learning; semisupervised learning; SELF-REPRESENTATION; IMAGE FEATURES; SCALE;
D O I
10.1109/TNNLS.2023.3278379
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multiview clustering has become a research hotspot in recent years due to its excellent capability of heterogeneous data fusion. Although a great deal of related works has appeared one after another, most of them generally overlook the potentials of prior knowledge utilization and progressive sample learning, resulting in unsatisfactory clustering performance in real-world applications. To deal with the aforementioned drawbacks, in this article, we propose a semisupervised progressive representation learning approach for deep multiview clustering (namely, SPDMC). Specifically, to make full use of the discriminative information contained in prior knowledge, we design a flexible and unified regularization, which models the sample pairwise relationship by enforcing the learned view-specific representation of must-link (ML) samples (cannot-link (CL) samples) to be similar (dissimilar) with cosine similarity. Moreover, we introduce the self-paced learning (SPL) paradigm and take good care of two characteristics in terms of both complexity and diversity when progressively learning multiview representations, such that the complementarity across multiple views can be squeezed thoroughly. Through comprehensive experiments on eight widely used image datasets, we prove that the proposed approach can perform better than the state-of-the-art opponents.
引用
收藏
页码:14341 / 14355
页数:15
相关论文
共 81 条
  • [1] Akaho S., 2006, ARXIV
  • [2] Andrew R., 2013, INT C MACH LEARN, P1247
  • [3] [Anonymous], 1967, P 5 BERKELEY S MATH, DOI DOI 10.1007/S11665-016-2173-6
  • [4] [Anonymous], 2003, INT JOINT C ARTIFICI
  • [5] Semi-Supervised Clustering With Constraints of Different Types From Multiple Information Sources
    Bai, Liang
    Liang, JiYe
    Cao, Fuyuan
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (09) : 3247 - 3258
  • [6] Benton A, 2019, 4TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP (REPL4NLP-2019), P1
  • [7] Bishop C. M., 2007, Pattern Recognition and Machine Learning Information Science and Statistics, V1st
  • [8] Structural Deep Clustering Network
    Bo, Deyu
    Wang, Xiao
    Shi, Chuan
    Zhu, Meiqi
    Lu, Emiao
    Cui, Peng
    [J]. WEB CONFERENCE 2020: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2020), 2020, : 1400 - 1410
  • [9] Cai, 2013, P 23 INT JOINT C ART, P2598, DOI DOI 10.5555/2540128.2540503
  • [10] Heterogeneous Image Features Integration via Multi-Modal Semi-Supervised Learning Model
    Cai, Xiao
    Nie, Feiping
    Cai, Weidong
    Huang, Heng
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 1737 - 1744