Regularized Semi-Supervised Metric Learning with Latent Structure Preserved

被引:1
作者
Wang, Qianying [1 ]
Lu, Ming [2 ]
Li, Meng [1 ]
Guan, Fei [1 ]
机构
[1] Hebei Univ Econ & Business, Coll Math & Stat, Shijiazhuang 050000, Hebei, Peoples R China
[2] Hebei Normal Univ, Sch Math Sci, Shijiazhuang 050000, Hebei, Peoples R China
关键词
Semi-supervised learning; metric learning; similarity; latent structure; clustering random forest;
D O I
10.1142/S1469026821500139
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Metric learning is a critical problem in classification. Most classifiers are based on a metric, the simplest one is the KNN classifier, whose outcome is directly decided by the given metric. This paper will discuss semi-supervised metric learning. Most traditional semi-supervised metric learning algorithms preserve the local structure of all the samples (including labeled and unlabeled) in the input space, when making the same labeled samples together and separating different labeled samples. In most existing methods, the local structure is calculated by the Euclidean distance which uses all the features. As we all know, high dimensional data lies on a low dimension manifold, and not all the features are discriminative. Thus, in this paper, we try to explore the latent structure of the samples and use the more discriminative features to calculate the local structure. The latent structure is learned by clustering random forest and cast into similarity between samples. Based on the hierarchical structure of the trees and the split function, the similarity is obtained from discriminant features. Experimental results on public data sets show our algorithm outperforms the traditional similar related algorithms.
引用
收藏
页数:19
相关论文
共 39 条
[1]  
[Anonymous], 2006, P NIPS
[2]  
[Anonymous], 2004, P 21 INT C MACH LEAR
[3]  
[Anonymous], 2004, ADV NEURAL INF PROCE
[4]  
[Anonymous], 2010, Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, July 25-28, 2010, DOI DOI 10.1145/1835804.183594
[5]  
Baghshah MS, 2009, 21ST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI-09), PROCEEDINGS, P1217
[6]   A PAC-style model for learning from labeled and unlabeled data [J].
Balcan, MF ;
Blum, A .
LEARNING THEORY, PROCEEDINGS, 2005, 3559 :111-126
[7]   Regularization and semi-supervised learning on large graphs [J].
Belkin, M ;
Matveeva, I ;
Niyogi, P .
LEARNING THEORY, PROCEEDINGS, 2004, 3120 :624-638
[8]  
Blum A., 1998, Proceedings of the Eleventh Annual Conference on Computational Learning Theory, P92, DOI 10.1145/279943.279962
[9]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[10]  
Breiman L. I., 1984, Encycl. Ecol., V40, P582