Dropout training for SVMs with data augmentation

被引:0
作者
Ning Chen
Jun Zhu
Jianfei Chen
Ting Chen
机构
[1] Tsinghua University,MOE Key lab of Bioinformatics, Bioinformatics Division and Center for Synthetic and Systems Biology, TNLIST
[2] Tsinghua University,State Key Lab of Intelligent Technology and Systems, Department of Computer Science and Technology
来源
Frontiers of Computer Science | 2018年 / 12卷
关键词
dropout; SVMs; logistic regression; data augmentation; iteratively reweighted least square;
D O I
暂无
中图分类号
学科分类号
摘要
Dropout and other feature noising schemes have shown promise in controlling over-fitting by artificially corrupting the training data. Though extensive studies have been performed for generalized linear models, little has been done for support vector machines (SVMs), one of the most successful approaches for supervised learning. This paper presents dropout training for both linear SVMs and the nonlinear extension with latent representation learning. For linear SVMs, to deal with the intractable expectation of the non-smooth hinge loss under corrupting distributions, we develop an iteratively re-weighted least square (IRLS) algorithm by exploring data augmentation techniques. Our algorithm iteratively minimizes the expectation of a reweighted least square problem, where the re-weights are analytically updated. For nonlinear latent SVMs, we consider learning one layer of latent representations in SVMs and extend the data augmentation technique in conjunction with first-order Taylor-expansion to deal with the intractable expected hinge loss and the nonlinearity of latent representations. Finally, we apply the similar data augmentation ideas to develop a new IRLS algorithm for the expected logistic loss under corrupting distributions, and we further develop a non-linear extension of logistic regression by incorporating one layer of latent representations. Our algorithms offer insights on the connection and difference between the hinge loss and logistic loss in dropout training. Empirical results on several real datasets demonstrate the effectiveness of dropout training on significantly boosting the classification accuracy of both linear and nonlinear SVMs.
引用
收藏
页码:694 / 713
页数:19
相关论文
共 55 条
  • [1] Srivastava N(2014)Dropout: a simple way to prevent neural networks from overfitting Journal of Machine Learning Research 15 1929-1958
  • [2] Hinton G(2011)Data augmentation for support vector machines Bayesian Analysis 6 1-24
  • [3] Krizhevsky A(2013)Bayesian inference for logistic models using Polya-Gamma latent variables Journal of the American Statistical Association 108 1339-1349
  • [4] Sutskever I(2004)Are loss functions all the same Neural Computation 16 1063-1076
  • [5] Salakhutdinov R(2014)The dropout learning algorithm Artificial Intelligence 210 78-122
  • [6] Polson N G(2014)Dropout: a simple way to prevent neural networks from overfitting Journal of Machine Learning Research 15 1929-1958
  • [7] Scott S L(2014)Semi-supervised and unsupervised extreme learning machines IEEE Transactions on Cybernetics 44 2405-2417
  • [8] Polson N G(2014)Follow the leader with dropout perturbations Proceedings of Machine Learning Research 35 949-974
  • [9] Scott J G(2015)On the inductive bias of dropout Journal of Machine Learning Research 16 3403-3454
  • [10] Windle J(1996)Mean field theory for sigmoid belief networks Journal of Artificial Intelligence Research 4 61-76