Kernel-driven similarity learning

被引:107
作者
Kang, Zhao [1 ,2 ]
Peng, Chong [2 ]
Cheng, Qiang [2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Sichuan, Peoples R China
[2] Southern Illinois Univ, Dept Comp Sci, Carbondale, IL 62901 USA
基金
美国国家科学基金会;
关键词
Similarity measure; Nonlinear relation; Sparse representation; Kernel method; Multiple kernel learning; Clustering; Recommender systems; SPARSE REPRESENTATION; SIGNAL RECOVERY;
D O I
10.1016/j.neucom.2017.06.005
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Similarity measure is fundamental to many machine learning and data mining algorithms. Predefined similarity metrics are often data-dependent and sensitive to noise. Recently, data-driven approach which learns similarity information from data has drawn significant attention. The idea is to represent a data point by a linear combination of all (other) data points. However, it is often the case that more complex relationships beyond linear dependencies exist in the data. Based on the well known fact that kernel trick can capture the nonlinear structure information, we extend this idea to kernel spaces. Nevertheless, such an extension brings up another issue: its algorithm performance is largely determined by the choice of kernel, which is often unknown in advance. Therefore, we further propose a multiple kernel-based learning method. By doing so, our model can learn both linear and nonlinear similarity information, and automatically choose the most suitable kernel. As a result, our model is capable of learning complete similarity information hidden in data set. Comprehensive experimental evaluations of our algorithms on clustering and recommender systems demonstrate its superior performance compared to other state-ofthe-art methods. This performance also shows the great potential of our proposed algorithm for other possible applications. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:210 / 219
页数:10
相关论文
共 55 条
[11]   Near-optimal signal recovery from random projections: Universal encoding strategies? [J].
Candes, Emmanuel J. ;
Tao, Terence .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2006, 52 (12) :5406-5425
[12]  
Chen SSB, 2001, SIAM REV, V43, P129, DOI [10.1137/S003614450037906X, 10.1137/S1064827596304010]
[13]   Moving beyond Linearity and Independence in Top-N Recommender Systems [J].
Christakopoulou, Evangelia .
PROCEEDINGS OF THE 8TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'14), 2014, :409-412
[14]  
de Sousa Celso Andre R., 2013, Machine Learning and Knowledge Discovery in Databases. European Conference, ECML PKDD 2013. Proceedings: LNCS 8190, P160, DOI 10.1007/978-3-642-40994-3_11
[15]   Item-based top-N recommendation algorithms [J].
Deshpande, M ;
Karypis, G .
ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2004, 22 (01) :143-177
[16]   For most large underdetermined systems of linear equations the minimal l1-norm solution is also the sparsest solution [J].
Donoho, DL .
COMMUNICATIONS ON PURE AND APPLIED MATHEMATICS, 2006, 59 (06) :797-829
[17]  
Du L, 2015, PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), P3476
[18]   Image denoising via sparse and redundant representations over learned dictionaries [J].
Elad, Michael ;
Aharon, Michal .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2006, 15 (12) :3736-3745
[19]  
Elhamifar Ehsan, 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), P2790, DOI 10.1109/CVPRW.2009.5206547
[20]   Sparse Representation With Kernels [J].
Gao, Shenghua ;
Tsang, Ivor Wai-Hung ;
Chia, Liang-Tien .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2013, 22 (02) :423-434