Finite-Time Convergence Rates of Decentralized Markovian Stochastic Approximation

被引:0
作者
Wang, Pengfei [1 ,2 ,3 ]
Zheng, Nenggan [2 ,3 ,4 ,5 ,6 ]
机构
[1] Zhejiang Univ, Sch Software Technol, Ningbo, Peoples R China
[2] Zhejiang Univ, Qiushi Acad Adv Studies, Hangzhou, Peoples R China
[3] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou, Peoples R China
[4] Zhejiang Univ, State Key Lab Brain Machine Intelligence, Hangzhou, Peoples R China
[5] MOE, CCAI, Hangzhou, Peoples R China
[6] Zhejiang Prov Govt ZJU, Hangzhou, Peoples R China
来源
PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024 | 2024年
基金
中国国家自然科学基金;
关键词
OPTIMIZATION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Markovian stochastic approximation has recently aroused a great deal of interest in many fields; however, it is not well understood in decentralized settings. Decentralized Markovian stochastic approximation is far more challenging than its single-agent counterpart due to the complex coupling structure between decentralized communication and Markovian noise-corrupted local updates. In this paper, a decentralized local markovian stochastic approximation (DLMSA) algorithm has been proposed and attains a near-optimal convergence rate. Specifically, we first provide a local variant of decentralized Markovian stochastic approximation so that each agent performs multiple local updates and then periodically communicate with its neighbors. Furthermore, we propose DLMSA with compressed communication (C-DLMSA) for further reducing the communication overhead. In this way, each agent only needs to communicate compressed information (e.g., sign compression) with its neighbors. We show that C-DLMSA enjoys the same convergence rate as that of the original DLMSA. Finally, we verify our theoretical results by applying our methods to solve multi-task reinforcement learning problems over multi-agent systems.
引用
收藏
页码:5082 / 5090
页数:9
相关论文
共 51 条
  • [1] Alacaoglu A, 2023, PR MACH LEARN RES, V202, P458
  • [2] [Anonymous], 2022, IEEE Transactions on Pattern Analysis and Machine Intelligence
  • [3] Benveniste Albert, 2012, Adaptive algorithms and stochastic approximations, V22
  • [4] Bertsekas Dimitri, 1996, Neuro-dynamic programming, DOI DOI 10.1137/1022042
  • [5] Bhandari Jalaj, 2018, PMLR, P1691
  • [6] Borkar V. S., 2009, STOCHASTIC APPROXIMA, V48
  • [7] Borkar V, 2024, Arxiv, DOI arXiv:2110.14427
  • [8] Optimization Methods for Large-Scale Machine Learning
    Bottou, Leon
    Curtis, Frank E.
    Nocedal, Jorge
    [J]. SIAM REVIEW, 2018, 60 (02) : 223 - 311
  • [9] Chen TY, 2021, Arxiv, DOI arXiv:1812.03239
  • [10] Chen ZW, 2021, Arxiv, DOI arXiv:2102.01567