Sample Out-of-Sample Inference Based on Wasserstein Distance

被引:4
作者
Blanchet, Jose [1 ]
Kang, Yang [2 ]
机构
[1] Stanford Univ, Dept Management Sci & Engn, Stanford, CA 94305 USA
[2] Columbia Univ, Dept Stat, New York, NY 10027 USA
关键词
nonparametric statistics; probability; distributionally robust optimization; optimal transport; Wasserstein distance; EMPIRICAL LIKELIHOOD; CONFIDENCE-INTERVALS; ESTIMATING EQUATIONS; RATIO; OPTIMIZATION; BANDS;
D O I
10.1287/opre.2020.2028
中图分类号
C93 [管理学];
学科分类号
12 ; 1201 ; 1202 ; 120202 ;
摘要
We present a novel inference approach that we call sample out-of-sample inference. The approach can be used widely, ranging from semisupervised learning to stress testing, and it is fundamental in the application of data-driven distributionally robust optimization. Our method enables measuring the impact of plausible out-of-sample scenarios in a given performance measure of interest, such as a financial loss. The methodology is inspired by empirical likelihood (EL), but we optimize the empirical Wasserstein distance (instead of the empirical likelihood) induced by observations. From a methodological standpoint, our analysis of the asymptotic behavior of the induced Wasserstein-distance profile function shows dramatic qualitative differences relative to EL. For instance, in contrast to EL, which typically yields chi-squared weak convergence limits, our asymptotic distributions are often not chi-squared. Also, the rates of convergence that we obtain have some dependence on the dimension in a nontrivial way but remain controlled as the dimension increases.
引用
收藏
页码:985 / 1013
页数:29
相关论文
共 56 条