Accelerating Communication-Efficient Federated Multi-Task Learning With Personalization and Fairness

被引:1
作者
Xie, Renyou [1 ]
Li, Chaojie [1 ]
Zhou, Xiaojun [2 ]
Dong, Zhaoyang [3 ]
机构
[1] Univ New South Wales, Sch Elect Engn & Telecommun, Sydney, NSW 2052, Australia
[2] Cent South Univ, Sch Automat, Changsha 410083, Peoples R China
[3] City Univ Hong Kong, Dept Elect Engn, Hong Kong, Peoples R China
基金
中国国家自然科学基金; 澳大利亚研究理事会;
关键词
Data models; Convergence; Internet of Things; Costs; Training; Task analysis; Multitasking; Communication efficiency; data heterogeneity; federated learning; local momentum technique; multi-task learning; CLIENT SELECTION; ALGORITHM;
D O I
10.1109/TPDS.2024.3411815
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Federated learning techniques provide a promising framework for collaboratively training a machine learning model without sharing users' data, and delivering a security solution to guarantee privacy during the model training of IoT devices. Nonetheless, challenges posed by data heterogeneity and communication resource constraints make it difficult to develop an efficient federated learning algorithm in terms of the low order of convergence rate. It could significantly deteriorate the quality of service for critical machine learning tasks, e.g., facial recognition, which requires an edge-ready, low-power, low-latency training algorithm. To address these challenges, a communication-efficient federated learning approach is proposed in this paper where the momentum technique is leveraged to accelerate the convergence rate while largely reducing the communication requirements. First, a federated multi-task learning framework by which the learning tasks are reformulated by the multi-objective optimization problem is introduced to address the data heterogeneity. The multiple gradient descent algorithm is harnessed to find the common gradient descending direction for all participants so that the common features can be learned and no sacrifice on each clients' performance. Second, to reduce communication costs, a local momentum technique with global information is developed to speed up the convergence rate, where the convergence analysis of the proposed method under non-convex case is studied. It is proved that the proposed local momentum can actually achieve the same acceleration as the global momentum, whereas it is more robust than algorithms that solely rely on the acceleration by the global momentum. Third, the generalization of the proposed acceleration approach is investigated which is demonstrated by the accelerated variation of FedAvg. Finally, the performance of the proposed method on the learning model accuracy, convergence rate, and robustness to data heterogeneity, is investigated by empirical experiments on four public datasets, while a real-world IoT platform is constructed to demonstrate the communication efficiency of the proposed method.
引用
收藏
页码:2239 / 2253
页数:15
相关论文
共 50 条
[1]  
Ackerman Evan., 2016, IEEE Spectr., V53, P14
[2]  
[Anonymous], 2009, Cifar-10
[3]   Deep learning and face recognition: the state of the art [J].
Balaban, Stephen .
BIOMETRIC AND SURVEILLANCE TECHNOLOGY FOR HUMAN AND ACTIVITY IDENTIFICATION XII, 2015, 9457
[4]  
Cohen G, 2017, IEEE IJCNN, P2921, DOI 10.1109/IJCNN.2017.7966217
[5]  
Das R, 2022, PR MACH LEARN RES, V180, P496
[6]   A fast and elitist multiobjective genetic algorithm: NSGA-II [J].
Deb, K ;
Pratap, A ;
Agarwal, S ;
Meyarivan, T .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2002, 6 (02) :182-197
[7]   Multiple-gradient descent algorithm (MGDA) for multiobjective optimization [J].
Desideri, Jean-Antoine .
COMPTES RENDUS MATHEMATIQUE, 2012, 350 (5-6) :313-318
[8]   Dermatologist-level classification of skin cancer with deep neural networks [J].
Esteva, Andre ;
Kuprel, Brett ;
Novoa, Roberto A. ;
Ko, Justin ;
Swetter, Susan M. ;
Blau, Helen M. ;
Thrun, Sebastian .
NATURE, 2017, 542 (7639) :115-+
[9]  
Finn C, 2017, PR MACH LEARN RES, V70
[10]   Steepest descent methods for multicriteria optimization [J].
Fliege, J ;
Svaiter, BF .
MATHEMATICAL METHODS OF OPERATIONS RESEARCH, 2000, 51 (03) :479-494