An Efficient Algorithm for a Class of Large-Scale Support Vector Machines Exploiting Hidden Sparsity

被引:3
作者
Niu, Dunbiao [1 ]
Wang, Chengjing [2 ,3 ]
Tang, Peipei [4 ]
Wang, Qingsong [5 ,6 ]
Song, Enbin [7 ,8 ]
机构
[1] Sichuan Univ, Coll Math, Chengdu 610065, Peoples R China
[2] Southwest Jiaotong Univ, Sch Math, Chengdu 611756, Peoples R China
[3] Southwest Jiaotong Univ, Natl Engn Lab Integrated Transportat Big Data Appl, Chengdu 611756, Peoples R China
[4] Zhejiang Univ City Coll, Sch Comp & Comp Sci, Hangzhou 310015, Peoples R China
[5] Beihang Univ, Sch Math Sci, Beijing 100191, Peoples R China
[6] Southwest Jiaotong Univ, Sch Math, Chengdu 611756, Peoples R China
[7] Sichuan Univ, Coll Math, Chengdu 610064, Peoples R China
[8] Sichuan Univ, Sch Aeronaut & Astronaut, Chengdu 610064, Peoples R China
基金
中国国家自然科学基金;
关键词
Support vector machines; semismooth Newton method; augmented Lagrangian method; hidden sparsity; QUADRATIC-PROGRAMMING PROBLEMS; SELECTION; NEWTON;
D O I
10.1109/TSP.2022.3221837
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Support vector machines (SVMs) are successful supervised learning models that analyze data for classification and regression. Previous work has demonstrated the superiority of the SVMs in dealing with the high dimensional, low sample size problems. However, the increase of the sample size brings great challenges to accurately and efficiently solve the large-scale SVMs, especially for the nonlinear kernel SVMs, which may lead to huge computational costs and unaffordable storage burden. In this paper, we propose a highly efficient sparse semismooth Newton (SsN) based augmented Lagrangian (AL) method for solving a class of large-scale SVMs that can be formulated as a convex quadratic programming problem with a linear equality constraint and a simple box constraint. The asymptotic superlinear convergence rate of both the primal and the dual iteration sequences generated by the AL method is guaranteed due to the piecewise linear-quadratic structure of the problem. Furthermore, we reveal the close connection between the number of support vectors and the sparse structure of the generalized Jacobian for the inner subproblem of the AL method. By exploiting this hidden sparsity, the inner subproblem can be solved by the SsN method efficiently and accurately, which greatly reduces the storage burden and computational costs. In particular, for the nonlinear kernel SVMs, since the sparse structure may not manifest in the early iterations of the AL method, we solve a linear kernel SVM approximated by the random Fourier features method to produce a good initial point, and then transfer to solve the original problem. Numerical experiments demonstrate that the proposed algorithm outperforms the current state-of-the-art solvers for the large-scale SVMs.
引用
收藏
页码:5608 / 5623
页数:16
相关论文
共 56 条
[1]  
Abe S, 2002, NEURAL NETWORKS FOR SIGNAL PROCESSING XII, PROCEEDINGS, P89, DOI 10.1109/NNSP.2002.1030020
[2]  
Abe S., 2010, ADV PTRN RECOGNIT
[3]  
[Anonymous], 1999, Advances in kernel methods: Support vector learning
[4]  
[Anonymous], Lectures on convex optimization
[5]  
[Anonymous], 2007, PROC 24 INT C MACH L
[6]  
Bach F., 2002, Learning with kernels: support vector machines, regularization, optimization, and beyond
[7]   PROJECTED GRADIENT METHODS FOR LINEARLY CONSTRAINED PROBLEMS [J].
CALAMAI, PH ;
MORE, JJ .
MATHEMATICAL PROGRAMMING, 1987, 39 (01) :93-116
[8]   LIBSVM: A Library for Support Vector Machines [J].
Chang, Chih-Chung ;
Lin, Chih-Jen .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2011, 2 (03)
[9]  
Chang KW, 2008, J MACH LEARN RES, V9, P1369
[10]   A study on SMO-type decomposition methods for support vector machines [J].
Chen, Pai-Hsuen ;
Fan, Rong-En ;
Lin, Chih-Jen .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2006, 17 (04) :893-908