共 40 条
Communication-Efficient Federated Learning: A Second Order Newton-Type Method With Analog Over-the-Air Aggregation
被引:11
作者:
Krouka, Mounssif
[1
]
Elgabli, Anis
[1
]
Ben Issaid, Chaouki
[1
]
Bennis, Mehdi
[1
]
机构:
[1] Univ Oulu, Ctr Wireless Commun, Oulu 90014, Finland
来源:
IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING
|
2022年
/
6卷
/
03期
基金:
芬兰科学院;
关键词:
Convergence;
Training;
Privacy;
Data models;
Convex functions;
Collaborative work;
Atmospheric modeling;
Distributed optimization;
communication-efficient federated learning;
second-order methods;
analog-over-the-air aggregation;
ADMM;
STOCHASTIC GRADIENT DESCENT;
EDGE;
D O I:
10.1109/TGCN.2022.3173420
中图分类号:
TN [电子技术、通信技术];
学科分类号:
0809 ;
摘要:
Owing to their fast convergence, second-order Newton-type learning methods have recently received attention in the federated learning (FL) setting. However, current solutions are based on communicating the Hessian matrices from the devices to the parameter server, at every iteration, incurring a large number of communication rounds; calling for novel communication-efficient Newton-type learning methods. In this article, we propose a novel second-order Newton-type method that, similarly to its first-order counterpart, requires every device to share only a model-sized vector at each iteration while hiding the gradient and Hessian information. In doing so, the proposed approach is significantly more communication-efficient and privacy-preserving. Furthermore, by leveraging the over-the-air aggregation principle, our method inherits privacy guarantees and obtains much higher communication efficiency gains. In particular, we formulate the problem of learning the inverse Hessian-gradient product as a quadratic problem that is solved in a distributed way. The framework alternates between updating the inverse Hessian-gradient product using a few alternating direction method of multipliers (ADMM) steps, and updating the global model using Newton's method. Numerical results show that our proposed approach is more communication-efficient and scalable under noisy channels for different scenarios and across multiple datasets.
引用
收藏
页码:1862 / 1874
页数:13
相关论文
共 40 条