Distributionally Robust Optimization and Generalization in Kernel Methods

被引:0
作者
Staib, Matthew [1 ]
Jegelka, Stefanie [1 ]
机构
[1] MIT, CSAIL, Cambridge, MA 02139 USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019) | 2019年 / 32卷
关键词
CONVERGENCE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Distributionally robust optimization (DRO) has attracted attention in machine learning due to its connections to regularization, generalization, and robustness. Existing work has considered uncertainty sets based on phi-divergences and Wasserstein distances, each of which have drawbacks. In this paper, we study DRO with uncertainty sets measured via maximum mean discrepancy (MMD). We show that MMD DRO is roughly equivalent to regularization by the Hilbert norm and, as a byproduct, reveal deep connections to classic results in statistical learning. In particular, we obtain an alternative proof of a generalization bound for Gaussian kernel ridge regression via a DRO lense. The proof also suggests a new regularizer. Our results apply beyond kernel methods: we derive a generically applicable approximation of MMD DRO, and show that it generalizes recent work on variance-based regularization.
引用
收藏
页数:11
相关论文
共 42 条
  • [1] [Anonymous], ARXIV170507152
  • [2] [Anonymous], 2022, MATH OPER RES
  • [3] [Anonymous], 2016, P MACHINE LEARNING R
  • [4] [Anonymous], 2018, ARXIV180410556
  • [5] Robust Solutions of Optimization Problems Affected by Uncertain Probabilities
    Ben-Tal, Aharon
    den Hertog, Dick
    De Waegenaere, Anja
    Melenberg, Bertrand
    Rennen, Gijs
    [J]. MANAGEMENT SCIENCE, 2013, 59 (02) : 341 - 357
  • [6] Data-driven robust optimization
    Bertsimas, Dimitris
    Gupta, Vishal
    Kallus, Nathan
    [J]. MATHEMATICAL PROGRAMMING, 2018, 167 (02) : 235 - 292
  • [7] Bietti A., 2019, PR MACH LEARN RES, P664
  • [8] Bietti A, 2019, J MACH LEARN RES, V20
  • [9] Blanchet J., 2016, J APPL PROBAB
  • [10] Chwialkowski K, 2016, PR MACH LEARN RES, V48