Wasserstein Embedding Learning for Deep Clustering: A Generative Approach

被引:19
作者
Cai, Jinyu [1 ]
Zhang, Yunhe [1 ]
Wang, Shiping [1 ]
Fan, Jicong [2 ,3 ]
Guo, Wenzhong [1 ]
机构
[1] Fuzhou Univ, Coll Comp & Data Sci, Fujian 350108, Peoples R China
[2] Chinese Univ Hong Kong, Shenzhen 518172, Peoples R China
[3] Shenzhen Res Inst Big Data, Shenzhen 518172, Peoples R China
基金
中国国家自然科学基金;
关键词
Training; Data models; Generative adversarial networks; Clustering methods; Task analysis; Deep learning; Decoding; Unsupervised learning; clustering analysis; Wasserstein embedding; generative models; auto-encoder; ADVERSARIAL NETWORKS; IDENTIFICATION; SELECTION;
D O I
10.1109/TMM.2024.3369862
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning-based clustering methods, especially those incorporating deep generative models, have recently shown noticeable improvement on many multimedia benchmark datasets. However, existing generative models still suffer from unstable training, and the gradient vanishes, which results in the inability to learn desirable embedded features for clustering. In this paper, we aim to tackle this problem by exploring the capability of Wasserstein embedding in learning representative embedded features and introducing a new clustering module for jointly optimizing embedding learning and clustering. To this end, we propose Wasserstein embedding clustering (WEC), which integrates robust generative models with clustering. By directly minimizing the discrepancy between the prior and marginal distribution, we transform the optimization problem of Wasserstein distance from the original data space into embedding space, which differs from other generative approaches that optimize in the original data space. Consequently, it naturally allows us to construct a joint optimization framework with the designed clustering module in the embedding layer. Due to the substitutability of the penalty term in Wasserstein embedding, we further propose two types of deep clustering models by selecting different penalty terms. Comparative experiments conducted on nine publicly available multimedia datasets with several state-of-the-art methods demonstrate the effectiveness of our method.
引用
收藏
页码:7567 / 7580
页数:14
相关论文
共 65 条
[1]  
Arjovsky M, 2017, PR MACH LEARN RES, V70
[2]   Hybrid dimension reduction by integrating feature selection with feature extraction method for text clustering [J].
Bharti, Kusum Kumari ;
Singh, Pramod Kumar .
EXPERT SYSTEMS WITH APPLICATIONS, 2015, 42 (06) :3105-3114
[3]  
Borghuis V, 2020, PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P5225
[4]   Unsupervised deep clustering via contractive feature representation and focal loss [J].
Cai, Jinyu ;
Wang, Shiping ;
Xu, Chaoyang ;
Guo, Wenzhong .
PATTERN RECOGNITION, 2022, 123
[5]   Deep Self-Evolution Clustering [J].
Chang, Jianlong ;
Meng, Gaofeng ;
Wang, Lingfeng ;
Xiang, Shiming ;
Pan, Chunhong .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (04) :809-823
[6]   Self-Paced Enhanced Low-Rank Tensor Kernelized Multi-View Subspace Clustering [J].
Chen, Yongyong ;
Wang, Shuqin ;
Xiao, Xiaolin ;
Liu, Youfa ;
Hua, Zhongyun ;
Zhou, Yicong .
IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 :4054-4066
[7]   Mean shift: A robust approach toward feature space analysis [J].
Comaniciu, D ;
Meer, P .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2002, 24 (05) :603-619
[8]  
Dilokthanakul N, 2017, PROC INT C LEARN REP, P1
[9]   Balanced Self-Paced Learning for Generative Adversarial Clustering Network [J].
Dizaji, Kamran Ghasedi ;
Wang, Xiaoqian ;
Deng, Cheng ;
Huang, Heng .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4386-4395
[10]   Deep Clustering via Joint Convolutional Autoencoder Embedding and Relative Entropy Minimization [J].
Dizaji, Kamran Ghasedi ;
Herandi, Amirhossein ;
Deng, Cheng ;
Cai, Weidong ;
Huang, Heng .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :5747-5756