Soft Margin Support Vector Classification as Buffered Probability Minimization

被引:0
作者
Norton, Matthew [1 ]
Mafusalov, Alexander [1 ]
Uryasev, Stan [1 ]
机构
[1] Univ Florida, Dept Ind & Syst Engn, Risk Management & Financial Engn Lab, Gainesville, FL 32611 USA
关键词
Support Vector Machines; Buffered Probability of Exceedance; Conditional Value-at-Risk; Binary Classification; Robust Optimization; ALGORITHMS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we show that the popular C-SVM, soft-margin support vector classifier is equivalent to minimization of Buffered Probability of Exceedance (bPOE), a recently introduced characterization of uncertainty. To show this, we introduce a new SVM formulation, called the EC-SVM, which is derived from a simple bPOE minimization problem that is easy to interpret with a meaningful free parameter, optimal objective value, and probabilistic derivation. Over the range of its free parameter, the EC-SVM has both a convex and non-convex case which we connect to existing SVM formulations. We first show that the C-SVM, formulated with any regularization norm, is equivalent to the convex EC-SVM. Similarly, we show that the E nu-SVM is equivalent to the EC-SVM over its entire parameter range, which includes both the convex and non-convex case. These equivalences, coupled with the interpretability of the EC-SVM, allow us to gain surprising new insights into the C-SVM and fully connect soft margin support vector classification with superquantile and bPOE concepts. We also show that the EC-SVM can easily be cast as a robust optimization problem, where bPOE is minimized with data lying in a fixed uncertainty set. This reformulation allows us to clearly differentiate between the convex and non-convex case, with convexity associated with pessimistic views of uncertainty and non-convexity associated with optimistic views of uncertainty. Finally, we address some practical considerations. First, we show that these new insights can assist in making parameter selection more efficient. Second, we discuss optimization approaches for solving the EC-SVM. Third, we address the issue of generalization, providing generalization bounds for both bPOE and misclassification rate.
引用
收藏
页码:1 / 43
页数:43
相关论文
共 24 条
  • [1] [Anonymous], 2008, P 25 INT C MACH LEAR
  • [2] Bi J., 2005, Adv. Neural Inf. Process. Syst., V17, P161
  • [3] Stability and generalization
    Bousquet, O
    Elisseeff, A
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2002, 2 (03) : 499 - 526
  • [4] Distributed optimization and statistical learning via the alternating direction method of multipliers
    Boyd S.
    Parikh N.
    Chu E.
    Peleato B.
    Eckstein J.
    [J]. Foundations and Trends in Machine Learning, 2010, 3 (01): : 1 - 122
  • [5] Training ν-support vector classifiers:: Theory and algorithms
    Chang, CC
    Lin, CJ
    [J]. NEURAL COMPUTATION, 2001, 13 (09) : 2119 - 2147
  • [6] Training a support vector machine in the primal
    Chapelle, Olivier
    [J]. NEURAL COMPUTATION, 2007, 19 (05) : 1155 - 1178
  • [7] CORTES C, 1995, MACH LEARN, V20, P273, DOI 10.1023/A:1022627411411
  • [8] Davis Justin, 2014, 20144 U FLOR
  • [9] Mafusalov A., 2015, 20141 U FLOR
  • [10] Norton M., 2014, 20142 U FLOR