A Unified Scheme for Distance Metric Learning and Clustering via Rank-Reduced Regression

被引:10
作者
Guo, Wenzhong [1 ,2 ]
Shi, Yiqing [1 ,2 ]
Wang, Shiping [1 ,2 ]
机构
[1] Fuzhou Univ, Coll Math & Comp Sci, Fuzhou 350116, Peoples R China
[2] Fuzhou Univ, Fujian Prov Key Lab Network Comp & Intelligent In, Fuzhou 350116, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2021年 / 51卷 / 08期
基金
中国国家自然科学基金;
关键词
Measurement; Reliability; Optimization; Clustering algorithms; Covariance matrices; Task analysis; Convergence; Clustering; distance metric learning; machine learning; matrix factorization; rank-reduced regression; ALGORITHM;
D O I
10.1109/TSMC.2019.2946398
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distance metric learning aims to learn a positive semidefinite matrix such that similar samples are preserved with small distances while dissimilar ones are mapped with big values above a predefined margin. It can facilitate to improve the performance of certain learning tasks. In this article, distance metric learning and clustering are integrated into an unified framework via rank-reduced regression. First, distance metric learning is proved to be consistent with rank-reduced regression, which provides a new perspective to learn structured regularization matrices. Second, orthogonal and non-negative rank-reduced regression problems are addressed individually for clustering, and the corresponding algorithms with proved convergence are proposed. Finally, both distance metric learning and clustering are addressed simultaneously in the problem formulation, which may trigger some new insights for learning an effective clustering oriented low-dimensional embedding. To show the superior performance of the proposed method, we compare it with several state-of-the-art clustering approaches. And, extensive experiments on the test datasets demonstrate the superiority of the proposed method.
引用
收藏
页码:5218 / 5229
页数:12
相关论文
共 65 条
[1]   ESTIMATING LINEAR RESTRICTIONS ON REGRESSION COEFFICIENTS FOR MULTIVARIATE NORMAL DISTRIBUTIONS [J].
ANDERSON, TW .
ANNALS OF MATHEMATICAL STATISTICS, 1951, 22 (03) :327-351
[2]  
[Anonymous], 2002, Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining, DOI DOI 10.1145/775047.775110
[3]   An approach to supervised distance metric learning based on difference of convex functions programming [J].
Bac Nguyen ;
De Baets, Bernard .
PATTERN RECOGNITION, 2018, 81 :562-574
[4]   Data Clustering and Graph Partitioning via Simulated Mixing [J].
Bhatti, Shahzad ;
Beck, Carolyn ;
Nedic, Angelia .
IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2019, 6 (03) :253-266
[5]  
Bilen H, 2015, PROC CVPR IEEE, P1081, DOI 10.1109/CVPR.2015.7298711
[6]  
Bilenko M., 2004, P 21 INT C MACH LEAR, P1
[7]   Graph Regularized Nonnegative Matrix Factorization for Data Representation [J].
Cai, Deng ;
He, Xiaofei ;
Han, Jiawei ;
Huang, Thomas S. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2011, 33 (08) :1548-1560
[8]  
Calzada A, 2013, INT CONF MACH LEARN, P1836, DOI 10.1109/ICMLC.2013.6890895
[9]   Reduced rank regression via adaptive nuclear norm penalization [J].
Chen, Kun ;
Dong, Hongbo ;
Chan, Kung-Sik .
BIOMETRIKA, 2013, 100 (04) :901-920
[10]   DHeat: A Density Heat-Based Algorithm for Clustering With Effective Radius [J].
Chen, Yewang ;
Tang, Shengyu ;
Pei, Songwen ;
Wang, Cheng ;
Du, Jixiang ;
Xiong, Naixue .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2018, 48 (04) :649-660