A PAC-Bayesian margin bound for linear classifiers: Why SVMs work

被引:0
|
作者
Herbrich, R [1 ]
Graepel, T [1 ]
机构
[1] Tech Univ Berlin, Dept Comp Sci, Stat Res Grp, Berlin, Germany
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 13 | 2001年 / 13卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a bound on the generalisation error of linear classifiers in terms of a refined margin quantity on the training set. The result is obtained in a PAC-Bayesian framework and is based on geometrical arguments in the space of linear classifiers. The new bound constitutes an exponential improvement of the so far tightest margin bound by Shawe-Taylor et al. [8] and scales logarithmically in the inverse margin. Even in the case of less training examples than input dimensions sufficiently large margins lead to non-trivial bound values and - for maximum margins - to a vanishing complexity term. Furthermore, the classical margin is too coarse a measure for the essential quantity that controls the generalisation error: the volume ratio between the whole hypothesis space and the subset of consistent hypotheses. The practical relevance of the result lies in the fact that the well-known support vector machine is optimal w.r.t. the new bound only if the feature vectors are all of the same length. As a consequence we recommend to use SVMs on normalised feature vectors only - a recommendation that is well supported by our numerical experiments on two benchmark data sets.
引用
收藏
页码:224 / 230
页数:7
相关论文
共 12 条
  • [1] A PAC-Bayesian margin bound for linear classifiers
    Herbrich, R
    Graepel, T
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2002, 48 (12) : 3140 - 3150
  • [2] Simplified PAC-Bayesian margin bounds
    McAllester, D
    LEARNING THEORY AND KERNEL MACHINES, 2003, 2777 : 203 - 215
  • [3] A PAC-Bayesian Bound for Lifelong Learning
    Pentina, Anastasia
    Lampert, Christoph H.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 991 - 999
  • [4] Some new PAC-Bayesian bounds and their use in selection of regularization parameter for linear SVMs
    Sahu, Puja
    Hemachandra, Nandyala
    PROCEEDINGS OF THE ACM INDIA JOINT INTERNATIONAL CONFERENCE ON DATA SCIENCE AND MANAGEMENT OF DATA (CODS-COMAD'18), 2018, : 240 - 248
  • [5] PAC-Bayesian Bounds on Rate-Efficient Classifiers
    Abbas, Alhabib
    Andreopoulos, Yiannis
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 1 - 9
  • [6] PAC-Bayesian Bound for the Conditional Value at Risk
    Mhammedi, Zakaria
    Guedj, Benjamin
    Williamson, Robert C.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [7] A PAC-Bayesian Generalization Bound for Equivariant Networks
    Behboodi, Arash
    Cesa, Gabriele
    Cohen, Taco
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [8] Improved PAC-Bayesian Bounds for Linear Regression
    Shalaeva, Vera
    Esfahani, Alireza Fakhrizadeh
    Germain, Pascal
    Petreczky, Mihaly
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 5660 - 5667
  • [9] Improving Robust Generalization by Direct PAC-Bayesian Bound Minimization
    Wang, Zifan
    Ding, Nan
    Levinboim, Tomer
    Chen, Xi
    Soricut, Radu
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 16458 - 16468
  • [10] A Column Generation Bound Minimization Approach with PAC-Bayesian Generalization Guarantees
    Roy, Jean-Francis
    Marchand, Mario
    Laviolette, Francois
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 51, 2016, 51 : 1241 - 1249