Very Sparse LSSVM Reductions for Large-Scale Data

被引:75
作者
Mall, Raghvendra [1 ]
Suykens, Johan A. K. [1 ]
机构
[1] Katholieke Univ Leuven, Dept Elect Engn, Stadius Ctr Dynam Syst Signal Proc & Data Analyt, B-3001 Leuven, Belgium
基金
欧洲研究理事会;
关键词
L-0-norm; least squares support vector machine (LSSVM) classification and regression; reduced models; sparsity; SUPPORT VECTOR MACHINE; CROSS-VALIDATION; CHOICE;
D O I
10.1109/TNNLS.2014.2333879
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Least squares support vector machines (LSSVMs) have been widely applied for classification and regression with comparable performance with SVMs. The LSSVM model lacks sparsity and is unable to handle large-scale data due to computational and memory constraints. A primal fixed-size LSSVM(PFS-LSSVM) introduce sparsity using Nystrom approximation with a set of prototype vectors (PVs). The PFS-LSSVM model solves an overdetermined system of linear equations in the primal. However, this solution is not the sparsest. We investigate the sparsity-error tradeoff by introducing a second level of sparsity. This is done by means of L-0-norm-based reductions by iteratively sparsifying LSSVM and PFS-LSSVM models. The exact choice of the cardinality for the initial PV set is not important then as the final model is highly sparse. The proposed method overcomes the problem of memory constraints and high computational costs resulting in highly sparse reductions to LSSVM models. The approximations of the two models allow to scale the models to large-scale datasets. Experiments on real-world classification and regression data sets from the UCI repository illustrate that these approaches achieve sparse models without a significant tradeoff in errors.
引用
收藏
页码:1086 / 1097
页数:12
相关论文
共 34 条
[1]  
Baker C. T. H., 1983, NUMERICAL TREATMENT
[2]  
Blake C., UCI Repository of machine learning databases
[3]  
BOWMAN AW, 1984, BIOMETRIKA, V71, P353
[4]  
BURGES CJC, 1996, P 13 INT C MACH LEAR, P71
[5]   Enhancing Sparsity by Reweighted l1 Minimization [J].
Candes, Emmanuel J. ;
Wakin, Michael B. ;
Boyd, Stephen P. .
JOURNAL OF FOURIER ANALYSIS AND APPLICATIONS, 2008, 14 (5-6) :877-905
[6]   Improved sparse least-squares support vector machines [J].
Cawley, GC ;
Talbot, NLC .
NEUROCOMPUTING, 2002, 48 :1025-1031
[7]   LIBSVM: A Library for Support Vector Machines [J].
Chang, Chih-Chung ;
Lin, Chih-Jen .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2011, 2 (03)
[8]  
Cherkassky V, 1997, IEEE Trans Neural Netw, V8, P1564, DOI 10.1109/TNN.1997.641482
[9]   Optimized fixed-size kernel models for large data sets [J].
De Brabanter, K. ;
De Brabanter, J. ;
Suykens, J. A. K. ;
De Moor, B. .
COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2010, 54 (06) :1484-1504
[10]   Exact simplification of support vector solutions [J].
Downs, T ;
Gates, KE ;
Masters, A .
JOURNAL OF MACHINE LEARNING RESEARCH, 2002, 2 (02) :293-297