Semi-discrete optimal transport: hardness, regularization and numerical solution

被引:0
作者
Bahar Taşkesen
Soroosh Shafieezadeh-Abadeh
Daniel Kuhn
机构
[1] EPFL Lausanne,Risk Analytics and Optimization Chair
[2] Tepper School of Business,undefined
[3] CMU,undefined
来源
Mathematical Programming | 2023年 / 199卷
关键词
Optimal transport; Wasserstein distance; Complexity; P-hardness; Discrete choice models; Distributionally robust optimization; Stochastic gradient descent algorithms;
D O I
暂无
中图分类号
学科分类号
摘要
Semi-discrete optimal transport problems, which evaluate the Wasserstein distance between a discrete and a generic (possibly non-discrete) probability measure, are believed to be computationally hard. Even though such problems are ubiquitous in statistics, machine learning and computer vision, however, this perception has not yet received a theoretical justification. To fill this gap, we prove that computing the Wasserstein distance between a discrete probability measure supported on two points and the Lebesgue measure on the standard hypercube is already #\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\#$$\end{document}P-hard. This insight prompts us to seek approximate solutions for semi-discrete optimal transport problems. We thus perturb the underlying transportation cost with an additive disturbance governed by an ambiguous probability distribution, and we introduce a distributionally robust dual optimal transport problem whose objective function is smoothed with the most adverse disturbance distributions from within a given ambiguity set. We further show that smoothing the dual objective function is equivalent to regularizing the primal objective function, and we identify several ambiguity sets that give rise to several known and new regularization schemes. As a byproduct, we discover an intimate relation between semi-discrete optimal transport problems and discrete choice models traditionally studied in psychology and economics. To solve the regularized optimal transport problems efficiently, we use a stochastic gradient descent algorithm with imprecise stochastic gradient oracles. A new convergence analysis reveals that this algorithm improves the best known convergence guarantee for semi-discrete optimal transport problems with entropic regularizers.
引用
收藏
页码:1033 / 1106
页数:73
相关论文
共 232 条
  • [1] Ahipaşaoğlu SD(2016)On the flexibility of using marginal distribution choice models in traffic equilibrium Transportation Research Part B: Methodological 91 130-158
  • [2] Arıkan U(2018)A convex optimization approach for computing correlated choice probabilities with many alternatives IEEE Trans. Autom. Control 64 190-205
  • [3] Natarajan K(2022)Asymptotics for semidiscrete entropic optimal transport SIAM J. Math. Anal. 54 1718-1741
  • [4] Ahipaşaoğlu SD(1988)A representative consumer theory of the logit model Int. Econ. Rev. 29 461-466
  • [5] Li X(1998)Minkowski-type theorems and least-squares clustering Algorithmica 20 61-76
  • [6] Natarajan K(2010)Self-concordant analysis for logistic regression Electronic J. Stat. 4 384-414
  • [7] Altschuler JM(2014)Adaptivity of averaged stochastic gradient descent to local strong convexity for logistic regression J. Mach. Learn. Res. 15 595-627
  • [8] Niles-Weed J(2000)A computational fluid mechanics solution to the Monge-Kantorovich mass transfer problem Numer. Math. 84 375-393
  • [9] Stromme AJ(2015)Iterative Bregman projections for regularized transportation problems SIAM J. Sci. Comput. 37 A1111-A1138
  • [10] Anderson SP(1981)A new algorithm for the assignment problem Math. Program. 21 152-171