Solution Path for Pin-SVM Classifiers With Positive and Negative τ Values

被引:25
作者
Huang, Xiaolin [1 ,2 ]
Shi, Lei [3 ]
Suykens, Johan A. K. [1 ]
机构
[1] Univ Leuven, Dept Elect Engn, Stadius Ctr Dynam Syst Signal Proc & Data Analyt, B-3001 Leuven, Belgium
[2] Shanghai Jiao Tong Univ, Inst Image Proc & Pattern Recognit, Shanghai 200400, Peoples R China
[3] Fudan Univ, Sch Math Sci, Shanghai Key Lab Contemporary Appl Math, Shanghai 200433, Peoples R China
基金
欧洲研究理事会; 中国国家自然科学基金;
关键词
Piecewise linear; pinball loss; solution path; support vector machine (SVM); SUPPORT VECTOR MACHINES; LOGISTIC-REGRESSION; LEAST ANGLE; CLASSIFICATION; ALGORITHM;
D O I
10.1109/TNNLS.2016.2547324
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Applying the pinball loss in a support vector machine (SVM) classifier results in pin-SVM. The pinball loss is characterized by a parameter tau. Its value is related to the quantile level and different tau values are suitable for different problems. In this paper, we establish an algorithm to find the entire solution path for pin-SVM with different tau values. This algorithm is based on the fact that the optimal solution to pin-SVM is continuous and piecewise linear with respect to tau. We also show that the nonnegativity constraint on tau is not necessary, i.e., tau can be extended to negative values. First, in some applications, a negative tau leads to better accuracy. Second, tau = -1 corresponds to a simple solution that links SVM and the classical kernel rule. The solution for tau = -1 can be obtained directly and then be used as a starting point of the solution path. The proposed method efficiently traverses tau values through the solution path, and then achieves good performance by a suitable tau. In particular, tau = 0 corresponds to C-SVM, meaning that the traversal algorithm can output a result at least as good as C-SVM with respect to validation error.
引用
收藏
页码:1584 / 1593
页数:10
相关论文
共 46 条
[1]   On the kernel rule for function classification [J].
Abraham, C. ;
Biau, G. ;
Cadre, B. .
ANNALS OF THE INSTITUTE OF STATISTICAL MATHEMATICS, 2006, 58 (03) :619-633
[2]   One-bit compressed sensing with non-Gaussian measurements [J].
Ai, Albert ;
Lapanowski, Alex ;
Plan, Yaniv ;
Vershynin, Roman .
LINEAR ALGEBRA AND ITS APPLICATIONS, 2014, 441 :222-239
[3]   Convexity, classification, and risk bounds [J].
Bartlett, PL ;
Jordan, MI ;
McAuliffe, JD .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2006, 101 (473) :138-156
[4]   On the rate of convergence of regularized boosting classifiers [J].
Blanchard, G ;
Lugosi, G ;
Vayatis, N .
JOURNAL OF MACHINE LEARNING RESEARCH, 2004, 4 (05) :861-894
[5]   Statistical performance of support vector machines [J].
Blanchard, Gilles ;
Bousquet, Olivier ;
Massart, Pascal .
ANNALS OF STATISTICS, 2008, 36 (02) :489-531
[6]  
Bottou L, 2007, LARGE SCALE KERNEL M, V3, P301, DOI DOI 10.7551/MITPRESS/7496.003.0003
[7]   LIBSVM: A Library for Support Vector Machines [J].
Chang, Chih-Chung ;
Lin, Chih-Jen .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2011, 2 (03)
[8]  
Chang KW, 2008, J MACH LEARN RES, V9, P1369
[9]   Partitioning nominal attributes in decision trees [J].
Coppersmith, D ;
Hong, SJ ;
Hosking, JRM .
DATA MINING AND KNOWLEDGE DISCOVERY, 1999, 3 (02) :197-217
[10]  
CORTES C, 1995, MACH LEARN, V20, P273, DOI 10.1023/A:1022627411411