Distributed Nonparametric Regression under Communication Constraints

被引:0
作者
Zhu, Yuancheng [1 ]
Lafferty, John [2 ]
机构
[1] Univ Penn, Wharton Sch, Dept Stat, Philadelphia, PA 19104 USA
[2] Yale Univ, Dept Stat & Data Sci, New Haven, CT 06520 USA
来源
INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80 | 2018年 / 80卷
关键词
ASYMPTOTIC EQUIVALENCE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper studies the problem of nonparametric estimation of a smooth function with data distributed across multiple machines. We assume an independent sample from a white noise model is collected at each machine, and an estimator of the underlying true function needs to be constructed at a central machine. We place limits on the number of bits that each machine can use to transmit information to the central machine. Our results give both asymptotic lower bounds and matching upper bounds on the statistical risk under various settings. We identify three regimes, depending on the relationship among the number of machines, the size of data available at each machine, and the communication budget. When the communication budget is small, the statistical risk depends solely on this communication bottleneck, regardless of the sample size. In the regime where the communication budget is large, the classic mini-max risk in the non-distributed estimation setting is recovered. In an intermediate regime, the statistical risk depends on both the sample size and the communication budget.
引用
收藏
页数:9
相关论文
共 20 条
[1]  
[Anonymous], 2013, PMLR
[2]  
[Anonymous], 2015, ARXIV150905457
[3]  
[Anonymous], 2013, Advances in Neural Information Processing Systems
[4]   The CEO problem [J].
Berger, T ;
Zhang, Z ;
Viswanathan, H .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1996, 42 (03) :887-902
[5]  
BLANCHARD G, 2016, ARXIV161007487
[6]   Communication Lower Bounds for Statistical Estimation Problems via a Distributed Data Processing Inequality [J].
Braverman, Mark ;
Garg, Ankit ;
Ma, Tengyu ;
Nguyen, Huy L. ;
Woodruff, David P. .
STOC'16: PROCEEDINGS OF THE 48TH ANNUAL ACM SIGACT SYMPOSIUM ON THEORY OF COMPUTING, 2016, :1011-1020
[7]  
Brown LD, 1996, ANN STAT, V24, P2384
[8]  
Chakrabarti D, 2017, J MACH LEARN RES, V18, P1
[9]  
Diakonikolas I., 2017, Advances in Neural Information Processing Systems, P6394
[10]  
Fan J., 2017, ARXIV170206488