VARIANCE REDUCTION-BOOSTED BYZANTINE ROBUSTNESS IN DECENTRALIZED STOCHASTIC OPTIMIZATION

被引:6
作者
Peng, Jie [1 ]
Li, Weiyu [2 ]
Ling, Qing [1 ]
机构
[1] Sun Yat Sen Univ, Guangzhou, Peoples R China
[2] Harvard Univ, Cambridge, MA 02138 USA
来源
2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP) | 2022年
关键词
Decentralized stochastic optimization; Byzantine robustness; variance reduction; GRADIENT DESCENT;
D O I
10.1109/ICASSP43922.2022.9746340
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
We consider the Byzantine-robust decentralized stochastic optimization problem, where every agent periodically communicates with its neighbors to exchange the local models, and then updates its own local model by stochastic gradient descent. However, an unknown number of the agents are Byzantine, and perform adversarially during the optimization process. Few works have considered this challenging scenario, and an existing method termed DECEMBER is unable to simultaneously achieve linear convergence speed and small learning error due to the stochastic noise. To eliminate the negative effect of the stochastic noise, we introduce two variance reduction methods, stochastic average gradient algorithm (SAGA) and loopless stochastic variance-reduced gradient (LSVRG), to Byzantine-robust decentralized stochastic optimization. The two resulting methods, DECEMBER-SAGA and DECEMBER-LSVRG, enjoy both linear convergence speeds and small learning errors. Numerical experiments demonstrate their effectiveness.
引用
收藏
页码:4283 / 4287
页数:5
相关论文
共 32 条
[1]  
[Anonymous], 2015, PROC CVPR IEEE, DOI DOI 10.1109/CVPR.2015.7298801
[2]  
[Anonymous], 2016, The Journal of Machine Learning Research
[3]   Robust Distributed Consensus Using Total Variation [J].
Ben-Ameur, Walid ;
Bianchi, Pascal ;
Jakubowicz, Jeremie .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2016, 61 (06) :1550-1564
[4]  
Blanchard P, 2017, ADV NEUR IN, V30
[5]   Distributed Approximate Newton's Method Robust to Byzantine Attackers [J].
Cao, Xinyang ;
Lai, Lifeng .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2020, 68 :6011-6025
[6]   Distributed Gradient Descent Algorithm Robust to an Arbitrary Number of Byzantine Attackers [J].
Cao, Xinyang ;
Lai, Lifeng .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2019, 67 (22) :5850-5864
[7]   Distributed Learning in the Nonconvex World: From batch data to streaming and beyond [J].
Chang, Tsung-Hui ;
Hong, Mingyi ;
Wai, Hoi-To ;
Zhang, Xinwei ;
Lu, Songtao .
IEEE SIGNAL PROCESSING MAGAZINE, 2020, 37 (03) :26-38
[8]   Distributed statistical machine learning in adversarial settings: Byzantine gradient descent [J].
Chen, Yudong ;
Su, Lili ;
Xu, Jiaming .
Proceedings of the ACM on Measurement and Analysis of Computing Systems, 2017, 1 (02)
[9]  
Defazio A, 2014, ADV NEUR IN, V27
[10]   Asynchronous Stochastic Gradient Descent Over Decentralized Datasets [J].
Du, Yubo ;
You, Keyou .
IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2021, 8 (03) :1212-1224