Communication-efficient Algorithms for Distributed Stochastic Principal Component Analysis

被引:0
|
作者
Garber, Dan [1 ]
Shamir, Ohad [2 ]
Srebro, Nathan [3 ]
机构
[1] Technion Israel Inst Technol, Haifa, Israel
[2] Weizmann Inst Sci, Rehovot, Israel
[3] Toyota Technol Inst, Chicago, IL 60637 USA
来源
INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70 | 2017年 / 70卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the fundamental problem of Principal Component Analysis in a statistical distributed setting in which each machine out of m stores a sample of n points sampled i.i.d. from a single unknown distribution. We study algorithms for estimating the leading principal component of the population covariance matrix that are both communication-efficient and achieve estimation error of the order of the centralized ERM solution that uses all mn samples. On the negative side, we show that in contrast to results obtained for distributed estimation under convexity assumptions, for the PCA objective, simply averaging the local ERM solutions cannot guarantee error that is consistent with the centralized ERM. We show that this unfortunate phenomena can be remedied by performing a simple correction step which correlates between the individual solutions, and provides an estimator that is consistent with the centralized ERM for sufficiently-large n. We also introduce an iterative distributed algorithm that is applicable in any regime of n, which is based on distributed matrix-vector products. The algorithm gives significant acceleration in terms of communication rounds over previous distributed algorithms, in a wide regime of parameters.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Stochastic Communication-Efficient Distributed Algorithms for Solving Linear Algebraic Equations
    Gao, Xiaobin
    Liu, Ji
    Basar, Tamer
    2016 IEEE CONFERENCE ON CONTROL APPLICATIONS (CCA), 2016,
  • [2] Communication-efficient algorithms for decentralized and stochastic optimization
    Lan, Guanghui
    Lee, Soomin
    Zhou, Yi
    MATHEMATICAL PROGRAMMING, 2020, 180 (1-2) : 237 - 284
  • [3] Communication-efficient algorithms for decentralized and stochastic optimization
    Guanghui Lan
    Soomin Lee
    Yi Zhou
    Mathematical Programming, 2020, 180 : 237 - 284
  • [4] Communication Efficient Distributed Kernel Principal Component Analysis
    Balcan, Maria-Florina
    Liang, Yingyu
    Song, Le
    Woodruff, David
    Xie, Bo
    KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, : 725 - 734
  • [5] Local Stochastic ADMM for Communication-Efficient Distributed Learning
    ben Issaid, Chaouki
    Elgabli, Anis
    Bennis, Mehdi
    2022 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2022, : 1880 - 1885
  • [6] Communication-Efficient Stochastic Gradient Descent Ascent with Momentum Algorithms
    Zhang, Yihan
    Qiu, Meikang
    Gao, Hongchang
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 4602 - 4610
  • [7] Differentially Private and Communication-Efficient Distributed Nonconvex Optimization Algorithms
    Xie, Antai
    Yi, Xinlei
    Wang, Xiaofan
    Cao, Ming
    Ren, Xiaoqiang
    arXiv, 2023,
  • [8] A Review of Distributed Algorithms for Principal Component Analysis
    Wu, Sissi Xiaoxiao
    Wai, Hoi-To
    Li, Lin
    Scaglione, Anna
    PROCEEDINGS OF THE IEEE, 2018, 106 (08) : 1321 - 1340
  • [9] Communication-Efficient Distributed Learning via Sparse and Adaptive Stochastic Gradient
    Deng, Xiaoge
    Li, Dongsheng
    Sun, Tao
    Lu, Xicheng
    IEEE TRANSACTIONS ON BIG DATA, 2025, 11 (01) : 234 - 246
  • [10] Communication-Efficient Adam-Type Algorithms for Distributed Data Mining
    Xian, Wenhan
    Huang, Feihu
    Huang, Heng
    2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2022, : 1245 - 1250