Turbo-aggregate: Breaking the quadratic aggregation barrier in secure federated learning

被引:120
作者
So J. [1 ]
Güler B. [2 ]
Avestimehr A.S. [1 ]
机构
[1] The Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, 90089, CA
[2] The Department of Electrical and Computer Engineering, University of California at Riverside, Riverside, 92521, CA
来源
IEEE Journal on Selected Areas in Information Theory | 2021年 / 2卷 / 01期
关键词
Federated learning; Privacy-preserving machine learning; Secure aggregation;
D O I
10.1109/JSAIT.2021.3054610
中图分类号
学科分类号
摘要
Federated learning is a distributed framework for training machine learning models over the data residing at mobile devices, while protecting the privacy of individual users. A major bottleneck in scaling federated learning to a large number of users is the overhead of secure model aggregation across many users. In particular, the overhead of the state-of-the-art protocols for secure model aggregation grows quadratically with the number of users. In this article, we propose the first secure aggregation framework, named Turbo-Aggregate, that in a network with N users achieves a secure aggregation overhead of O(N log N), as opposed to O(N2 ), while tolerating up to a user dropout rate of 50%. Turbo-Aggregate employs a multi-group circular strategy for efficient model aggregation, and leverages additive secret sharing and novel coding techniques for injecting aggregation redundancy in order to handle user dropouts while guaranteeing user privacy. We experimentally demonstrate that Turbo-Aggregate achieves a total running time that grows almost linear in the number of users, and provides up to 40× speedup over the state-of-the-art protocols with up to N = 200 users. Our experiments also demonstrate the impact of model size and bandwidth on the performance of Turbo-Aggregate. © 2021 IEEE.
引用
收藏
页码:479 / 489
页数:10
相关论文
共 55 条
[1]  
McMahan H.B., Moore E., Ramage D., Hampson S., Arcas B.A.Y., Communication-efficient learning of deep networks from decentralized data, Proc. Int. Conf. Artif. Int. Stat. (AISTATS), pp. 1273-1282, (2017)
[2]  
Bonawitz K., Et al., Practical secure aggregation for federated learning on user-held data, Proc. Conf. Neural Inf. Process. Syst., (2016)
[3]  
Bonawitz K., Et al., Practical secure aggregation for privacy-preserving machine learning, Proc. ACM SIGSAC Conf. Comput. Commun. Security, pp. 1175-1191, (2017)
[4]  
Kairouz P., Et al., Advances and Open Problems in Federated Learning, (2019)
[5]  
Zhu L., Liu Z., Han S., Deep leakage from gradients, Proc. Adv. Neural Inf. Process. Syst., pp. 14774-14784, (2019)
[6]  
Wang Z., Song M., Zhang Z., Song Y., Wang Q., Qi H., Beyond inferring class representatives: User-level privacy leakage from federated learning, Proc. IEEE INFOCOM Conf. Comput. Commun., pp. 2512-2520, (2019)
[7]  
Geiping J., Bauermeister H., Droge H., Moeller M., Inverting Gradients—How Easy Is It to Break Privacy in Federated Learning?, (2020)
[8]  
Yang T., Et al., Applied Federated Learning: Improving Google Keyboard Query Suggestions, (2018)
[9]  
McMahan H.B., Ramage D., Talwar K., Zhang L., Learning differentially private recurrent language models, Proc. Int. Conf. Learn. Represent. (ICLR), pp. 1-14, (2018)
[10]  
Bonawitz K., Et al., Towards federated learning at scale: System design, Proc. 2nd SysML Conf., (2019)