Scaling Up Kernel SVM on Limited Resources: A Low-Rank Linearization Approach

被引:37
作者
Lan, Liang [1 ]
Wang, Zhuang [2 ,3 ]
Zhe, Shandian [4 ]
Cheng, Wei [5 ]
Wang, Jun [6 ]
Zhang, Kai [7 ]
机构
[1] Hong Kong Baptist Univ, Dept Comp Sci, Hong Kong, Peoples R China
[2] Siemens Corp Res, Princeton, NJ 08540 USA
[3] Facebook, Menlo Pk, CA 94025 USA
[4] Univ Utah, Sch Comp, Salt Lake City, UT USA
[5] NEC Labs Amer, Princeton, NJ 08550 USA
[6] East China Normal Univ, Sch Comp Sci & Software Engn, Shanghai 200062, Peoples R China
[7] Temple Univ, Dept Comp & Informat Sci, Philadelphia, PA 19122 USA
关键词
Large-scale learning; low-rank approximation; metric learning; support vector machine (SVM); NYSTROM METHOD; MATRIX;
D O I
10.1109/TNNLS.2018.2838140
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Kernel support vector machines (SVMs) deliver state-of-the-art results in many real-world nonlinear classification problems, but the computational cost can be quite demanding in order to maintain a large number of support vectors. Linear SVM, on the other hand, is highly scalable to large data but only suited for linearly separable problems. In this paper, we propose a novel approach called low-rank linearized SVMto scale up kernel SVM on limited resources. Our approach transforms a nonlinear SVM to a linear one via an approximate empirical kernel map computed from efficient kernel low-rank decompositions. We theoretically analyze the gap between the solutions of the approximate and optimal rank-k kernel map, which in turn provides guidance on the sampling scheme of the Nystrom approximation. Furthermore, we extend it to a semisupervised metric learning scenario in which partially labeled samples can be exploited to further improve the quality of the low-rank embedding. Our approach inherits rich representability of kernel SVM and high efficiency of linear SVM. Experimental results demonstrate that our approach is more robust and achieves a better tradeoff between model representability and scalability against state-of-the-art algorithms for large-scale SVMs.
引用
收藏
页码:369 / 378
页数:10
相关论文
共 45 条
  • [31] Sonnenburg S., 2010, ICML, P999
  • [32] Tsang IW, 2005, J MACH LEARN RES, V6, P363
  • [33] Wang Z, 2012, J MACH LEARN RES, V13, P3103
  • [34] Wang Zhuang, 2011, P 17 ACM SIGKDD INT, P24, DOI 10.1145/2020408.2020420
  • [35] Williams CKI, 2001, ADV NEUR IN, V13, P682
  • [36] Wu MR, 2006, J MACH LEARN RES, V7, P603
  • [37] Robust Kernel Low-Rank Representation
    Xiao, Shijie
    Tan, Mingkui
    Xu, Dong
    Dong, Zhao Yang
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (11) : 2268 - 2281
  • [38] Xing E. P., 2003, NIPS, P521
  • [39] Yang T., 2012, ADV NEURAL INFORM PR, P476
  • [40] Large Linear Classification When Data Cannot Fit in Memory
    Yu, Hsiang-Fu
    Hsieh, Cho-Jui
    Chang, Kai-Wei
    Lin, Chih-Jen
    [J]. ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2012, 5 (04)