A nonlinear least squares quasi-Newton strategy for LP-SVR hyper-parameters selection

被引:8
作者
Rivas-Perea, Pablo [1 ]
Cota-Ruiz, Juan [2 ]
Rosiles, Jose-Gerardo [3 ]
机构
[1] Baylor Univ, Dept Comp Sci, Waco, TX 76798 USA
[2] Autonomous Univ Ciudad Juarez UACJ, Dept Elect & Comp Engn, Ciudad Juarez 32310, Chihuahua, Mexico
[3] Sci Applicat Int Corp, El Paso, TX 79925 USA
关键词
Support vector regression; Hyper-parameters; Large-scale LP-SVR; SUPPORT VECTOR REGRESSION; MODEL SELECTION; PERFORMANCE-MEASURES; HYPERPARAMETERS; CRITERIA; SPACE;
D O I
10.1007/s13042-013-0153-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper studies the problem of hyper-parameters selection for a linear programming-based support vector machine for regression (LP-SVR). The proposed model is a generalized method that minimizes a linear-least squares problem using a globalization strategy, inexact computation of first order information, and an existing analytical method for estimating the initial point in the hyper-parameters space. The minimization problem consists of finding the set of hyper-parameters that minimizes any generalization error function for different problems. Particularly, this research explores the case of two-class, multi-class, and regression problems. Simulation results among standard data sets suggest that the algorithm achieves statistically insignificant variability when measuring the residual error; and when compared to other methods for hyper-parameters search, the proposed method produces the lowest root mean squared error in most cases. Experimental analysis suggests that the proposed approach is better suited for large-scale applications for the particular case of an LP-SVR. Moreover, due to its mathematical formulation, the proposed method can be extended in order to estimate any number of hyper-parameters.
引用
收藏
页码:579 / 597
页数:19
相关论文
共 71 条
[71]  
Zhang XQ, 2007, PROCEEDINGS OF 2007 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, P3261