Distributed Stochastic Optimization Under a General Variance Condition

被引:1
|
作者
Huang, Kun [1 ]
Li, Xiao [1 ]
Pu, Shi [1 ]
机构
[1] Chinese Univ Hong Kong, Sch Data Sci, Shenzhen CUHK Shenzhen, Shenzhen 518172, Peoples R China
基金
中国国家自然科学基金;
关键词
Optimization; Linear programming; Distributed databases; Gradient methods; Convergence; Complexity theory; Particle measurements; Distributed optimization; nonconvex optimization; stochastic optimization; LEARNING-BEHAVIOR; CONVERGENCE;
D O I
10.1109/TAC.2024.3393169
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distributed stochastic optimization has drawn great attention recently due to its effectiveness in solving large-scale machine learning problems. Although numerous algorithms have been proposed and successfully applied to general practical problems, their theoretical guarantees mainly rely on certain boundedness conditions on the stochastic gradients, varying from uniform boundedness to the relaxed growth condition. In addition, how to characterize the data heterogeneity among the agents and its impacts on the algorithmic performance remains challenging. In light of such motivations, we revisit the classical federated averaging algorithm (McMahan et al., 2017) as well as the more recent SCAFFOLD method (Karimireddy et al., 2020) for solving the distributed stochastic optimization problem and establish the convergence results under only a mild variance condition on the stochastic gradients for smooth nonconvex objective functions. Almost sure convergence to a stationary point is also established under the condition. Moreover, we discuss a more informative measurement for data heterogeneity as well as its implications.
引用
收藏
页码:6105 / 6120
页数:16
相关论文
共 50 条
  • [21] Convergence in High Probability of Distributed Stochastic Gradient Descent Algorithms
    Lu, Kaihong
    Wang, Hongxia
    Zhang, Huanshui
    Wang, Long
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (04) : 2189 - 2204
  • [22] Solving Stochastic Compositional Optimization is Nearly as Easy as Solving Stochastic Optimization
    Chen, Tianyi
    Sun, Yuejiao
    Yin, Wotao
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 : 4937 - 4948
  • [23] Balancing Rates and Variance via Adaptive Batch-Size for Stochastic Optimization Problems
    Gao, Zhan
    Koppel, Alec
    Ribeiro, Alejandro
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2022, 70 : 3693 - 3708
  • [24] Stochastic distributed learning with gradient quantization and double-variance reduction
    Horvath, Samuel
    Kovalev, Dmitry
    Mishchenko, Konstantin
    Richtarik, Peter
    Stich, Sebastian
    OPTIMIZATION METHODS & SOFTWARE, 2023, 38 (01) : 91 - 106
  • [25] Distributed Stochastic Constrained Composite Optimization Over Time-Varying Network With a Class of Communication Noise
    Yu, Zhan
    Ho, Daniel W. C.
    Yuan, Deming
    Liu, Jie
    IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (06) : 3561 - 3573
  • [26] Primal-Dual Methods for Large-Scale and Distributed Convex Optimization and Data Analytics
    Jakovetic, Dusan
    Bajovic, Dragana
    Xavier, Joao
    Moura, Jose M. F.
    PROCEEDINGS OF THE IEEE, 2020, 108 (11) : 1923 - 1938
  • [27] Local Stochastic Factored Gradient Descent for Distributed Quantum State Tomography
    Kim, Junhyung Lyle
    Toghani, Mohammad Taha
    Uribe, Cesar A.
    Kyrillidis, Anastasios
    IEEE CONTROL SYSTEMS LETTERS, 2022, 7 : 199 - 204
  • [28] A Distributed Algorithm for Convex Constrained Optimization Under Noise
    Chatzipanagiotis, Nikolaos
    Zavlanos, Michael M.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2016, 61 (09) : 2496 - 2511
  • [29] Distributed stochastic gradient tracking methods
    Pu, Shi
    Nedic, Angelia
    MATHEMATICAL PROGRAMMING, 2021, 187 (1-2) : 409 - 457
  • [30] Push-LSVRG-UP: Distributed Stochastic Optimization Over Unbalanced Directed Networks With Uncoordinated Triggered Probabilities
    Hu, Jinhui
    Chen, Guo
    Li, Huaqing
    Shen, Zixiang
    Zhang, Weidong
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2023, 10 (02): : 934 - 950