Distributed Stochastic Optimization Under a General Variance Condition

被引:1
|
作者
Huang, Kun [1 ]
Li, Xiao [1 ]
Pu, Shi [1 ]
机构
[1] Chinese Univ Hong Kong, Sch Data Sci, Shenzhen CUHK Shenzhen, Shenzhen 518172, Peoples R China
基金
中国国家自然科学基金;
关键词
Optimization; Linear programming; Distributed databases; Gradient methods; Convergence; Complexity theory; Particle measurements; Distributed optimization; nonconvex optimization; stochastic optimization; LEARNING-BEHAVIOR; CONVERGENCE;
D O I
10.1109/TAC.2024.3393169
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distributed stochastic optimization has drawn great attention recently due to its effectiveness in solving large-scale machine learning problems. Although numerous algorithms have been proposed and successfully applied to general practical problems, their theoretical guarantees mainly rely on certain boundedness conditions on the stochastic gradients, varying from uniform boundedness to the relaxed growth condition. In addition, how to characterize the data heterogeneity among the agents and its impacts on the algorithmic performance remains challenging. In light of such motivations, we revisit the classical federated averaging algorithm (McMahan et al., 2017) as well as the more recent SCAFFOLD method (Karimireddy et al., 2020) for solving the distributed stochastic optimization problem and establish the convergence results under only a mild variance condition on the stochastic gradients for smooth nonconvex objective functions. Almost sure convergence to a stationary point is also established under the condition. Moreover, we discuss a more informative measurement for data heterogeneity as well as its implications.
引用
收藏
页码:6105 / 6120
页数:16
相关论文
共 50 条
  • [1] Distributed Stochastic Projection-Free Algorithm for Constrained Optimization
    Jiang, Xia
    Zeng, Xianlin
    Xie, Lihua
    Sun, Jian
    Chen, Jie
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2025, 70 (04) : 2479 - 2494
  • [2] Distributed Stochastic Gradient Tracking Algorithm With Variance Reduction for Non-Convex Optimization
    Jiang, Xia
    Zeng, Xianlin
    Sun, Jian
    Chen, Jie
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (09) : 5310 - 5321
  • [3] Distributed Coupled Multiagent Stochastic Optimization
    Alghunaim, Sulaiman A.
    Sayed, Ali H.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2020, 65 (01) : 175 - 190
  • [4] Distributed Evolution Strategies for Black-Box Stochastic Optimization
    He, Xiaoyu
    Zheng, Zibin
    Chen, Chuan
    Zhou, Yuren
    Luo, Chuan
    Lin, Qingwei
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (12) : 3718 - 3731
  • [5] Stochastic Successive Convex Approximation for General Stochastic Optimization Problems
    Ye, Chencheng
    Cui, Ying
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2020, 9 (06) : 755 - 759
  • [6] Distributed Momentum-Based Frank-Wolfe Algorithm for Stochastic Optimization
    Hou, Jie
    Zeng, Xianlin
    Wang, Gang
    Sun, Jian
    Chen, Jie
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2023, 10 (03) : 685 - 699
  • [7] Decentralized Stochastic Optimization With Pairwise Constraints and Variance Reduction
    Han, Fei
    Cao, Xuanyu
    Gong, Yi
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2024, 72 : 1960 - 1973
  • [8] A Stochastic Second-Order Proximal Method for Distributed Optimization
    Qiu, Chenyang
    Zhu, Shanying
    Ou, Zichong
    Lu, Jie
    IEEE CONTROL SYSTEMS LETTERS, 2023, 7 : 1405 - 1410
  • [9] S-DIGing: A Stochastic Gradient Tracking Algorithm for Distributed Optimization
    Li, Huaqing
    Zheng, Lifeng
    Wang, Zheng
    Yan, Yu
    Feng, Liping
    Guo, Jing
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2022, 6 (01): : 53 - 65
  • [10] Distributed Stochastic Variance Reduced Gradient Methods by Sampling Extra Data with Replacement
    Lee, Jason D.
    Lin, Qihang
    Ma, Tengyu
    Yang, Tianbao
    JOURNAL OF MACHINE LEARNING RESEARCH, 2017, 18