A Convex Model for Support Vector Distance Metric Learning

被引:12
作者
Ruan, Yibang [1 ]
Xiao, Yanshan [1 ]
Hao, Zhifeng [1 ,2 ]
Liu, Bo [3 ]
机构
[1] Guangdong Univ Technol, Sch Comp, Guangzhou 510006, Peoples R China
[2] Foshan Univ, Sch Math & Big Data, Foshan 528000, Peoples R China
[3] Guangdong Univ Technol, Sch Automat, Guangzhou 510006, Peoples R China
关键词
Support vector machines; Measurement; Training; Support vector machine classification; Kernel; Predictive models; Data models; Distance metric learning (DML); k nearest neighbor; support vector classification; FACIAL EXPRESSION RECOGNITION; TEXT CLASSIFICATION; MACHINES; PERFORMANCE;
D O I
10.1109/TNNLS.2021.3053266
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Distance metric learning (DML) aims to learn a distance metric to process the data distribution. However, most of the existing methods are kNN DML methods and employ the kNN model to classify the test instances. The drawback of kNN DML is that all training instances need to be accessed and stored to classify the test instances, and the classification performance is influenced by the setting of the nearest neighbor number k. To solve these problems, there are several DML methods that employ the SVM model to classify the test instances. However, all of them are nonconvex and the convex support vector DML method has not been explicitly proposed. In this article, we propose a convex model for support vector DML (CSV-DML), which is capable of replacing the kNN model of DML with the SVM model. To make CSV-DML can use the most kernel functions of the existing SVM methods, a nonlinear mapping is used to map the original instances into a feature space. Since the explicit form of nonlinear mapped instances is unknown, the original instances are further transformed into the kernel form, which can be calculated explicitly. CSV-DML is constructed to work directly on the kernel-transformed instances. Specifically, we learn a specific Mahalanobis distance metric from the kernel-transformed training instances and train a DML-based separating hyperplane based on it. An iterated approach is formulated to optimize CSV-DML, which is based on generalized block coordinate descent and can converge to the global optimum. In CSV-DML, since the dimension of kernel-transformed instances is only related to the number of original training instances, we develop a novel parameter reduction scheme for reducing the feature dimension. Extensive experiments show that the proposed CSV-DML method outperforms the previous methods.
引用
收藏
页码:3533 / 3546
页数:14
相关论文
共 61 条
[1]  
[Anonymous], 2009, Advances in Neural Information Processing Systems
[2]  
[Anonymous], 2006, IEEE CVPR
[3]  
[Anonymous], 2011, P CVPR WORKSHOPS JUN, DOI DOI 10.1109/CVPRW.2011.5981788
[4]  
[Anonymous], 2011, Acm T. Intel. Syst. Tec., DOI [DOI 10.1145/1961189.1961199, 10.1145/1961189.1961199]
[5]   Evaluating Open-Universe Face Identification on the Web [J].
Becker, Brian C. ;
Ortiz, Enrique G. .
2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2013, :904-911
[6]   A tutorial on Support Vector Machines for pattern recognition [J].
Burges, CJC .
DATA MINING AND KNOWLEDGE DISCOVERY, 1998, 2 (02) :121-167
[7]   Support vector machines for histogram-based image classification [J].
Chapelle, O ;
Haffner, P ;
Vapnik, VN .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1999, 10 (05) :1055-1064
[8]  
Chen S., 2019, ADV NEUR IN, P4223
[9]  
Chen YX, 2004, J MACH LEARN RES, V5, P913
[10]   Facial Expression Recognition in JAFFE Dataset Based on Gaussian Process Classification [J].
Cheng, Fei ;
Yu, Jiangsheng ;
Xiong, Huilin .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2010, 21 (10) :1685-1690