Federated Learning via Over-the-Air Computation

被引:749
作者
Yang, Kai [1 ,2 ,3 ]
Jiang, Tao [1 ]
Shi, Yuanming [1 ]
Ding, Zhi [4 ]
机构
[1] ShanghaiTech Univ, Sch Informat Sci & Technol, Shanghai 201210, Peoples R China
[2] Chinese Acad Sci, Shanghai Inst Microsyst & Informat Technol, Shanghai 200050, Peoples R China
[3] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[4] Univ Calif Davis, Dept Elect & Comp Engn, Davis, CA 95616 USA
基金
美国国家科学基金会;
关键词
Federated learning; over-the-air computation; edge machine learning; sparse optimization; low-rank optimization; difference-of-convex-functions; DC programming; ANALOG FUNCTION COMPUTATION; OPTIMIZATION; SPARSE; ACCESS; DESIGN;
D O I
10.1109/TWC.2019.2961673
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The stringent requirements for low-latency and privacy of the emerging high-stake applications with intelligent devices such as drones and smart vehicles make the cloud computing inapplicable in these scenarios. Instead, edge machine learning becomes increasingly attractive for performing training and inference directly at network edges without sending data to a centralized data center. This stimulates a nascent field termed as federated learning for training a machine learning model on computation, storage, energy and bandwidth limited mobile devices in a distributed manner. To preserve data privacy and address the issues of unbalanced and non-IID data points across different devices, the federated averaging algorithm has been proposed for global model aggregation by computing the weighted average of locally updated model at each selected device. However, the limited communication bandwidth becomes the main bottleneck for aggregating the locally computed updates. We thus propose a novel over-the-air computation based approach for fast global model aggregation via exploring the superposition property of a wireless multiple-access channel. This is achieved by joint device selection and beamforming design, which is modeled as a sparse and low-rank optimization problem to support efficient algorithms design. To achieve this goal, we provide a difference-of-convex-functions (DC) representation for the sparse and low-rank function to enhance sparsity and accurately detect the fixed-rank constraint in the procedure of device selection. A DC algorithm is further developed to solve the resulting DC program with global convergence guarantees. The algorithmic advantages and admirable performance of the proposed methodologies are demonstrated through extensive numerical results.
引用
收藏
页码:2022 / 2035
页数:14
相关论文
共 55 条
[1]  
Abari Omid, 2015, 2015 IEEE Conference on Computer Communications (INFOCOM). Proceedings, P1742, DOI 10.1109/INFOCOM.2015.7218555
[2]   Robust Design for Massive CSI Acquisition in Analog Function Computation Networks [J].
Ang, Fan ;
Chen, Li ;
Zhao, Nan ;
Chen, Yunfei ;
Yu, F. Richard .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (03) :2361-2373
[3]  
[Anonymous], 2017, ADV NEURAL INFORM PR
[4]  
[Anonymous], 2014, Convex Optimiza- tion
[5]  
[Anonymous], 2018, ARXIV180900343
[6]  
[Anonymous], 2018, ARXIV181010167
[7]  
[Anonymous], 2017, ADV NEURAL INF PROC
[8]  
[Anonymous], IEEE T WIRELESS COMM
[9]   Adaptive Learning in Complex Reproducing Kernel Hilbert Spaces Employing Wirtinger's Subgradients [J].
Bouboulis, Pantelis ;
Slavakis, Konstantinos ;
Theodoridis, Sergios .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2012, 23 (03) :425-438
[10]   Multiuser MIMO Achievable Rates With Downlink Training and Channel State Feedback [J].
Caire, Giuseppe ;
Jindal, Nihar ;
Kobayashi, Mari ;
Ravindran, Niranjay .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2010, 56 (06) :2845-2866