Support vector machines and regularization

被引:0
作者
Cherkassky, V [1 ]
Ma, YQ [1 ]
机构
[1] Univ Minnesota, Dept Elect & Comp Engn, Minneapolis, MN 55455 USA
来源
Seventh IASTED International Conference on Signal and Image Processing | 2005年
关键词
function approximation; regularization; structural risk minimization;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, there has been a growing interest in Statistical Learning Theory, aka VC theory, due to many successful applications of Support Vector Machines (SVMs). Even though most theoretical results in VC-theory (including all main concepts underlying SVM methodology) have been developed over 25 years ago, these concepts are occasionally misunderstood in the research community. This paper compares standard SVM regression and the regularization for learning dependencies from data. We point out that SVM approach has been developed in VC-theory under risk minimization approach, whereas the regularization approach has been developed under function approximation setting. This distinction is especially important since regularization-based learning is often presented as a purely constructive methodology (with no clearly stated problem setting), even though original regularization theory has been introduced under clearly stated function approximation setting. Further, we present empirical comparisons illustrating the effect of different mechanisms for complexity control (i.e., epsilon-insensitive loss vs standard ridge regression) on the generalization performance, under very simple settings using synthetic data sets. These comparisons suggest that the SVM approach to complexity control (via epsilon-loss) is more appropriate for learning under sparse high-dimensional settings.
引用
收藏
页码:166 / 171
页数:6
相关论文
共 14 条
  • [1] [Anonymous], 1982, ESTIMATION DEPENDENC
  • [2] Risk bounds for model selection via penalization
    Barron, A
    Birgé, L
    Massart, P
    [J]. PROBABILITY THEORY AND RELATED FIELDS, 1999, 113 (03) : 301 - 413
  • [3] Practical selection of SVM parameters and noise estimation for SVM regression
    Cherkassky, V
    Ma, YQ
    [J]. NEURAL NETWORKS, 2004, 17 (01) : 113 - 126
  • [4] Cherkassky V, 2002, LECT NOTES COMPUT SC, V2415, P687
  • [5] Cherkassky V.S., 1998, LEARNING DATA CONCEP, V1st ed.
  • [6] Regularization networks and support vector machines
    Evgeniou, T
    Pontil, M
    Poggio, T
    [J]. ADVANCES IN COMPUTATIONAL MATHEMATICS, 2000, 13 (01) : 1 - 50
  • [7] Friedman J., 2001, The elements of statistical learning, V1, DOI DOI 10.1007/978-0-387-21606-5
  • [8] REGULARIZATION ALGORITHMS FOR LEARNING THAT ARE EQUIVALENT TO MULTILAYER NETWORKS
    POGGIO, T
    GIROSI, F
    [J]. SCIENCE, 1990, 247 (4945) : 978 - 982
  • [9] Smola A. J., 2002, Learning With Kernels: Support Vector Machines, Regularization, Optimization, and Beyond
  • [10] TIKHONOV N, 1963, DOKL AKAD NAUK SSSR, V153, P501