On the Structure, Covering, and Learning of Poisson Multinomial Distributions

被引:11
作者
Daskalakis, Constantinos [1 ]
Kamath, Gautam [1 ]
Tzamos, Christos [1 ]
机构
[1] MIT, Elect Engn & Comp Sci, 77 Massachusetts Ave, Cambridge, MA 02139 USA
来源
2015 IEEE 56TH ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE | 2015年
关键词
Structure; Learning; Applied probability; Gaussian distribution; Multivariate statistics; Limit theorem; NASH EQUILIBRIA;
D O I
10.1109/FOCS.2015.77
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
An (n, k)-Poisson Multinomial Distribution (PMD) is the distribution of the sum of n independent random vectors supported on the set B-k = {e(1), . . . , e(k)} of standard basis vectors in R-k. We prove a structural characterization of these distributions, showing that, for all epsilon > 0, any (n, k)-Poisson multinomial random vector is epsilon-close, in total variation distance, to the sum of a discretized multidimensional Gaussian and an independent (poly(k/epsilon), k)-Poisson multinomial random vector. Our structural characterization extends the multi-dimensional CLT of [2], by simultaneously applying to all approximation requirements epsilon. In particular, it overcomes factors depending on log n and, importantly, the minimum eigenvalue of the PMD's covariance matrix. We use our structural characterization to obtain an epsilon-cover, in total variation distance, of the set of all (n, k)-PMDs, significantly improving the cover size of [3], [4], and obtaining the same qualitative dependence of the cover size on n and epsilon as the k = 2 cover of [5], [6]. We further exploit this structure to show that (n, k)-PMDs can be learned to within e in total variation distance from (O) over tilde (k)(1/epsilon(2)) samples, which is near-optimal in terms of dependence on e and independent of n. In particular, our result generalizes the single-dimensional result of [7] for Poisson binomials to arbitrary dimension. Finally, as a corollary of our results on PMDs, we give a (O) over tilde k(1/epsilon(2)) sample algorithm for learning (n, k)-sums of independent integer random variables (SIIRVs), which is near-optimal for constant k.
引用
收藏
页码:1203 / 1217
页数:15
相关论文
共 19 条
[1]  
Acharya J, 2014, IEEE INT SYMP INFO, P1682, DOI 10.1109/ISIT.2014.6875120
[2]  
Acharya Jayadev., 2015, Proceedings of SODA, P1829, DOI DOI 10.1137/1.9781611973730.122
[3]  
[Anonymous], 1988, Journal of Appl. Probab.
[4]  
[Anonymous], THEORY PROBAB APPL
[5]   A Lyapunov-type bound in Rd [J].
Bentkus, V .
THEORY OF PROBABILITY AND ITS APPLICATIONS, 2004, 49 (02) :311-323
[6]  
Daskalakis C., 2014, PROBABILITY THEORY R
[7]  
Daskalakis C, 2012, STOC'12: PROCEEDINGS OF THE 2012 ACM SYMPOSIUM ON THEORY OF COMPUTING, P709
[8]   Approximate Nash equilibria in anonymous games [J].
Daskalakis, Constantinos ;
Papadimitriou, Christos H. .
JOURNAL OF ECONOMIC THEORY, 2015, 156 :207-245
[9]   Learning Sums of Independent Integer Random Variables [J].
Daskalakis, Constantinos ;
Diakonikolas, Ilias ;
O'Donnell, Ryan ;
Servedio, Rocco A. ;
Tan, Li-Yang .
2013 IEEE 54TH ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS), 2013, :217-226
[10]  
Daskalakis C, 2009, ACM S THEORY COMPUT, P75