Surrogate-assisted multi-objective model selection for support vector machines

被引:29
作者
Rosales-Perez, Alejandro [1 ]
Gonzalez, Jesus A. [1 ]
Coello Coello, Carlos A. [2 ]
Jair Escalante, Hugo [1 ]
Reyes-Garcia, Carlos A. [1 ]
机构
[1] Inst Nacl Astrofis Opt & Electr, Dept Comp Sci, Puebla 72840, Mexico
[2] IPN, Ctr Invest & Estudios Avanzados CINVESTAV, Dept Comp Sci, Evolutionary Computat Grp EVOCINV, Mexico City 07360, DF, Mexico
关键词
Model selection; Multi-objective optimization; Support vector machines; Surrogate-assisted optimization; VARIANCE ANALYSIS; BIAS; ALGORITHM; OPTIMIZATION; PARAMETERS; VALUES;
D O I
10.1016/j.neucom.2014.08.075
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Classification is one of the most well-known tasks in supervised learning. A vast number of algorithms for pattern classification have been proposed so far. Among these, support vector machines (SVMs) are one of the most popular approaches, due to the high performance reached by these methods in a wide number of pattern recognition applications. Nevertheless, the effectiveness of SVMs highly depends on their hyper-parameters. Besides the fine-tuning of their hyper-parameters, the way in which the features are scaled as well as the presence of non-relevant features could affect their generalization performance. This paper introduces an approach for addressing model selection for support vector machines used in classification tasks. In our formulation, a model can be composed of feature selection and pre-processing methods besides the SVM classifier. We formulate the model selection problem as a multi-objective one, aiming to minimize simultaneously two components that are closely related to the error of a model: bias and variance components, which are estimated in an experimental fashion. A surrogate-assisted evolutionary multi-objective optimization approach is adopted to explore the hyper-parameters space. We adopted this approach due to the fact that estimating the bias and variance could be computationally expensive. Therefore, by using surrogate-assisted optimization, we expect to reduce the number of solutions evaluated by the fitness functions so that the computational cost would also be reduced. Experimental results conducted on benchmark datasets widely used in the literature, indicate that highly competitive models with a fewer number of fitness function evaluations are obtained by our proposal when it is compared to state of the art model selection methods. (C) 2014 Elsevier B.V. All rights reserved.
引用
收藏
页码:163 / 172
页数:10
相关论文
共 47 条
[1]  
[Anonymous], TECHNICAL REPORT
[2]  
[Anonymous], 2002, Evolutionary Methods for Design, Optimization and Control with Application to Industrial Problems (EUROGEN 2001)
[3]  
[Anonymous], 2012, 2012 INT JOINT C NEU
[4]   Automatic model selection for the optimization of SVM kernels [J].
Ayat, NE ;
Cheriet, M ;
Suen, CY .
PATTERN RECOGNITION, 2005, 38 (10) :1733-1745
[5]   A multi-objective artificial immune algorithm for parameter optimization in support vector machine [J].
Aydin, Ilhan ;
Karakose, Mehmet ;
Akin, Erhan .
APPLIED SOFT COMPUTING, 2011, 11 (01) :120-129
[6]   A PSO and pattern search based memetic algorithm for SVMs parameters optimization [J].
Bao, Yukun ;
Hu, Zhongyi ;
Xiong, Tao .
NEUROCOMPUTING, 2013, 117 :98-106
[7]  
Bishop Christopher, 2006, Pattern Recognition and Machine Learning, DOI 10.1117/1.2819119
[8]  
Boser B. E., 1992, Proceedings of the Fifth Annual ACM Workshop on Computational Learning Theory, P144, DOI 10.1145/130385.130401
[9]  
Cawley GC, 2007, J MACH LEARN RES, V8, P841
[10]  
Cawley GC, 2010, J MACH LEARN RES, V11, P2079