Model selection for gaussian kernel support vector machines in random fourier feature space

被引:0
|
作者
Feng C. [1 ]
Liao S. [1 ]
机构
[1] School of Computer Science and Technology, Tianjin University, Tianjin
来源
Jisuanji Yanjiu yu Fazhan/Computer Research and Development | 2016年 / 53卷 / 09期
基金
中国国家自然科学基金;
关键词
Cross validation; Gaussian kernel; Model selection; Random Fourier features; Support vector machines (SVMs);
D O I
10.7544/issn1000-1239.2016.20150489
中图分类号
学科分类号
摘要
Model selection is very critical to support vector machines (SVMs). Standard SVMs typically suffer from cubic time complexity in data size since they solve the convex quadratic programming problems. However, it usually needs to train hundreds/thousands of SVMs for model selection, which is prohibitively time-consuming for medium-scale datasets and very difficult to scale up to large-scale problems. In this paper, by using random Fourier features to approximate Gaussian kernel, a novel and efficient approach to model selection of kernel SVMs is proposed. Firstly, the random Fourier feature mapping is used to embed the infinite-dimensional implicit feature space into an explicit random feature space. An error bound between the accurate model obtained by training kernel SVM and the approximate one returned by the linear SVM in the random feature space is derived. Then, in the random feature space, a model selection approach to kernel SVM is presented. Under the guarantee of the model error upper bound, by applying the linear SVMs in the random feature space to approximate the corresponding kernel SVMs, the approximate model selection criterion can be efficiently calculated and used to assess the relative goodness of the corresponding kernel SVMs. Finally, comparison experiments on benchmark datasets for cross validation model selection show the proposed approach can significantly improve the efficiency of model selection for kernel SVMs while guaranteeing test accuracy. © 2016, Science Press. All right reserved.
引用
收藏
页码:1971 / 1978
页数:7
相关论文
共 25 条
  • [1] Vapnik V.N., Statistical Learning Theory, (1998)
  • [2] Scholkopf B., Smola A.J., Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond, (2002)
  • [3] Chapelle O., Vapnik V., Model selection for support vector machines, Advances in Neural Information Processing Systems 12, pp. 230-236, (2000)
  • [4] Guyon I., Saffari A., Dror G., Et al., Model selection: Beyond the Bayesian/frequentist divide, Journal of Machine Learning Research, 11, pp. 61-87, (2010)
  • [5] Duan K., Keerthi S.S., Poo A.N., Evaluation of simple performance measures for tuning SVM hyperparameters, Neurocomputing, 51, pp. 41-59, (2003)
  • [6] Chapelle O., Vapnik V.N., Bousquet O., Et al., Choosing multiple parameters for support vector machines, Machine Learning, 46, 1-3, pp. 131-159, (2002)
  • [7] Vapnik V.N., Chapelle O., Bounds on error expectation for support vector machines, Neural Computation, 12, 9, pp. 2013-2036, (2000)
  • [8] Platt J.C., Fast Training of support vector machines using sequential minimal optimization, Advances in Kernel Methods: Support Vector Learning, pp. 185-208, (1999)
  • [9] Zhang T., Solving large scale linear prediction problems using stochastic gradient descent algorithms, Proc of the 21st Int Conf on Machine Learning, pp. 919-926, (2004)
  • [10] Joachims T., Training linear SVMs in linear time, Proc of the 12th ACM SIGKDD Int Conf on Knowledge Discovery and Data Mining, pp. 217-226, (2006)