Scalable Gaussian Kernel Support Vector Machines with Sublinear Training Time Complexity

被引:16
作者
Feng, Chang [1 ]
Liao, Shizhong [1 ]
机构
[1] Tianjin Univ, Sch Comp Sci & Technol, Tianjin 300350, Peoples R China
基金
中国国家自然科学基金;
关键词
Support Vector Machines (SVMs); Gaussian kernel; Random Fourier feature mapping; Subsampling; Parallel linear SVMs; Scalability; NYSTROM METHOD; MATRIX;
D O I
10.1016/j.ins.2017.08.033
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Gaussian kernel Support Vector Machines (SVMs) deliver state-of-the-art generalization performance for non-linear classification, but the time complexity of their training process is at least quadratic w.r.t the size of training set, preventing them from scaling up on large datasets. To address this issue, we propose a novel approach to large-scale kernel SVMs in sublinear time w.r.t. the size of training set, which combines three well-known and efficient techniques with theoretical guarantees. First, we subsample massive samples to reduce the sample size. Then, we use the random Fourier feature mapping on the sub samples to construct explicit random feature space where we can train a linear SVM to approximate the corresponding Gaussian kernel SVM. Finally, we use parallel algorithms to make our approach more scalable. Deriving the upper bounds of kernel matrix approximation error, hypothesis error and excess risk w.r.t. the size of training set and the dimension of random feature space, we establish the theoretical foundation of our proposed approach. In this way, we can reduce the time complexity of training kernel SVMs without sacrificing much accuracy. Our proposed approach achieves high accuracy and has sublinear training time complexity, which exhibits good scalability theoretically and empirically. (C) 2017 Elsevier Inc. All rights reserved.
引用
收藏
页码:480 / 494
页数:15
相关论文
共 49 条
[1]  
Afrati FN, 2006, LECT NOTES COMPUT SC, V3484, P1, DOI 10.1007/11671541_1
[2]  
[Anonymous], 2012, Scaling up Machine LearningParallel and Distributed Approaches
[3]  
[Anonymous], 2012, Proceedings of the fifteenth International Conference on Artificial Intelligence and Statistics
[4]  
[Anonymous], 2004, P 21 INT C MACH LEAR, DOI DOI 10.1145/1015330.1015332
[5]  
[Anonymous], P ADV NEUR INF PROC
[6]  
[Anonymous], 2011, FOURIER ANAL GROUPS
[7]  
[Anonymous], 2010, Journal of Machine Learning Research, DOI DOI 10.5555/1756006.1859899
[8]  
[Anonymous], 2012, Mining of massive datasets
[9]  
Bach F., 2013, PMLR, P185
[10]  
Bartlett P. L., 2003, Journal of Machine Learning Research, V3, P463, DOI 10.1162/153244303321897690