A Communication-Efficient Hierarchical Federated Learning Framework via Shaping Data Distribution at Edge

被引:17
作者
Deng, Yongheng [1 ]
Lyu, Feng [2 ]
Xia, Tengxi [1 ]
Zhou, Yuezhi [3 ]
Zhang, Yaoxue [1 ,3 ]
Ren, Ju [1 ,3 ]
Yang, Yuanyuan [4 ]
机构
[1] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol BNRist, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
[2] Cent South Univ, Sch Comp Sci & Engn, Changsha 410083, Peoples R China
[3] Zhongguancun Lab, Beijing 100084, Peoples R China
[4] SUNY Stony Brook, Dept Elect & Comp Engn, Stony Brook, NY 11794 USA
关键词
Costs; Data models; Servers; Computational modeling; Training data; Federated learning; Distributed databases; Hierarchical federated learning; communication efficiency; edge computing; distributed edge intelligence; RESOURCE-ALLOCATION;
D O I
10.1109/TNET.2024.3363916
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) enables collaborative model training over distributed computing nodes without sharing their privacy-sensitive raw data. However, in FL, iterative exchanges of model updates between distributed nodes and the cloud server can result in significant communication cost, especially when the data distributions at distributed nodes are imbalanced with requiring more rounds of iterations. In this paper, with our in-depth empirical studies, we disclose that extensive cloud aggregations can be avoided without compromising the learning accuracy if frequent aggregations can be enabled at edge network. To this end, we shed light on the hierarchical federated learning (HFL) framework, where a subset of distributed nodes can play as edge aggregators to support edge aggregations. Under the HFL framework, we formulate a communication cost minimization (CCM) problem to minimize the total communication cost required for model learning with a target accuracy by making decisions on edge aggragator selection and node-edge associations. Inspired by our data-driven insights that the potential of HFL lies in the data distribution at edge aggregators, we propose ShapeFL, i.e., SHaping dAta distRibution at Edge, to transform and solve the CCM problem. In ShapeFL, we divide the original problem into two sub-problems to minimize the per-round communication cost and maximize the data distribution diversity of edge aggregator data, respectively, and devise two light-weight algorithms to solve them accordingly. Extensive experiments are carried out based on several opened datasets and real-world network topologies, and the results demonstrate the efficacy of ShapeFL in terms of both learning accuracy and communication efficiency.
引用
收藏
页码:2600 / 2615
页数:16
相关论文
共 62 条
[1]  
Alistarh D, 2017, ADV NEUR IN, V30
[2]   Federated learning with hierarchical clustering of local updates to improve training on non-IID data [J].
Briggs, Christopher ;
Fan, Zhong ;
Andras, Peter .
2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
[3]   Space/Aerial-Assisted Computing Offloading for IoT Applications: A Learning-Based Approach [J].
Cheng, Nan ;
Lyu, Feng ;
Quan, Wei ;
Zhou, Conghao ;
He, Hongli ;
Shi, Weisen ;
Shen, Xuemin .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2019, 37 (05) :1117-1129
[4]  
Collins L, 2021, PR MACH LEARN RES, V139
[5]   Slashing Communication Traffic in Federated Learning by Transmitting Clustered Model Updates [J].
Cui, Laizhong ;
Su, Xiaoxin ;
Zhou, Yipeng ;
Pan, Yi .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (08) :2572-2589
[6]   Improving Federated Learning With Quality-Aware User Incentive and Auto-Weighted Model Aggregation [J].
Deng, Yongheng ;
Lyu, Feng ;
Ren, Ju ;
Chen, Yi-Chao ;
Yang, Peng ;
Zhou, Yuezhi ;
Zhang, Yaoxue .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (12) :4515-4529
[7]   AUCTION: Automated and Quality-Aware Client Selection Framework for Efficient Federated Learning [J].
Deng, Yongheng ;
Lyu, Feng ;
Ren, Ju ;
Wu, Huaqing ;
Zhou, Yuezhi ;
Zhang, Yaoxue ;
Shen, Xuemin .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (08) :1996-2009
[8]   SHARE: Shaping Data Distribution at Edge for Communication-Efficient Hierarchical Federated Learning [J].
Deng, Yongheng ;
Lyu, Feng ;
Ren, Ju ;
Zhang, Yongmin ;
Zhou, Yuezhi ;
Zhang, Yaoxue ;
Yang, Yuanyuan .
2021 IEEE 41ST INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2021), 2021, :24-34
[9]   Federated Learning Over Wireless Networks: Convergence Analysis and Resource Allocation [J].
Dinh, Canh T. ;
Tran, Nguyen H. ;
Nguyen, Minh N. H. ;
Hong, Choong Seon ;
Bao, Wei ;
Zomaya, Albert Y. ;
Gramoli, Vincent .
IEEE-ACM TRANSACTIONS ON NETWORKING, 2021, 29 (01) :398-409
[10]   Self-Balancing Federated Learning With Global Imbalanced Data in Mobile Systems [J].
Duan, Moming ;
Liu, Duo ;
Chen, Xianzhang ;
Liu, Renping ;
Tan, Yujuan ;
Liang, Liang .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2021, 32 (01) :59-71