Communication Efficient Federated Learning With Heterogeneous Structured Client Models

被引:7
|
作者
Hu, Yao [1 ,2 ]
Sun, Xiaoyan [3 ]
Tian, Ye [4 ,5 ]
Song, Linqi [1 ,2 ]
Tan, Kay Chen [6 ]
机构
[1] City Univ Hong Kong, Dept Comp Sci, Hong Kong 999077, Peoples R China
[2] City Univ Hong Kong, Shenzhen Res Inst, Shenzhen 518057, Peoples R China
[3] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou 221000, Jiangsu, Peoples R China
[4] Anhui Univ, Inst Phys Sci, Informat Mat & Intelligent Sensing Lab Anhui Prov, Hefei 230601, Peoples R China
[5] Anhui Univ, Inst Informat Technol, Informat Mat & Intelligent Sensing Lab Anhui Prov, Hefei 230601, Peoples R China
[6] Hong Kong Polytech Univ, Dept Comp, Hong Kong 999077, Peoples R China
来源
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE | 2023年 / 7卷 / 03期
基金
中国国家自然科学基金;
关键词
Servers; Costs; Matrix decomposition; Training; Data models; Optimization; Data privacy; Federated learning; heterogeneous structured model; neural network; singular value decomposition; FACTORIZATION; SYSTEMS;
D O I
10.1109/TETCI.2022.3209345
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Federated learning (FL) has recently attracted much attention due to its superior performance in privacy protection when processing data from different terminals. However, homogeneous deep learning models are pervasively adopted without considering the difference between distinct data in various clients, resulting in low learning performance and high communication costs. This paper thus proposes a novel FL framework with heterogeneous structured client models for handling different data scales and investigates its superiority over canonical FL with homogeneous models. Additionally, singular value decomposition is adopted on the client models to reduce the amount of transmitted data, i.e., the communication costs. The aggregation mechanism with multiple models on the central server is then presented based on the heterogeneous characteristics of the uploaded parameters and models. The proposed framework is applied to four benchmark classification datasets and a trend following task on electromagnetic radiation intensity time series data. Experimental results demonstrate that the proposed method can effectively improve the accuracy of local learning models and significantly reduce communication costs.
引用
收藏
页码:753 / 767
页数:15
相关论文
共 50 条
  • [1] Federated Learning With Heterogeneous Client Expectations: A Game Theory Approach
    Shen, Sheng
    Liu, Chi
    Lim, Teng Joon
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (12) : 8220 - 8237
  • [2] Client-Side Optimization Strategies for Communication-Efficient Federated Learning
    Mills, Jed
    Hu, Jia
    Min, Geyong
    IEEE COMMUNICATIONS MAGAZINE, 2022, 60 (07) : 60 - 66
  • [3] FedHe: Heterogeneous Models and Communication-Efficient Federated Learning
    Chan, Yun Hin
    Ngai, Edith C. H.
    2021 17TH INTERNATIONAL CONFERENCE ON MOBILITY, SENSING AND NETWORKING (MSN 2021), 2021, : 207 - 214
  • [4] Federated Learning in Heterogeneous Networks With Unreliable Communication
    Zheng, Paul
    Zhu, Yao
    Hu, Yulin
    Zhang, Zhengming
    Schmeink, Anke
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2024, 23 (04) : 3823 - 3838
  • [5] Efficient Client Sampling with Compression in Heterogeneous Federated Learning
    Marnissi, Ouiame
    El Hammouti, Hajar
    Bergou, El Houcine
    IEEE INFOCOM 2024-IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS, INFOCOM WKSHPS 2024, 2024,
  • [6] Heterogeneous Privacy Level-Based Client Selection for Hybrid Federated and Centralized Learning in Mobile Edge Computing
    Solat, Faranaksadat
    Patni, Sakshi
    Lim, Sunhwan
    Lee, Joohyung
    IEEE ACCESS, 2024, 12 : 108556 - 108572
  • [7] Compressed Client Selection for Efficient Communication in Federated Learning
    Mohamed, Aissa Hadj
    Assumpcao, Nicolas R. G.
    Astudillo, Carlos A.
    de Souza, Allan M.
    Bittencourt, Luiz F.
    Villas, Leandro A.
    2023 IEEE 20TH CONSUMER COMMUNICATIONS & NETWORKING CONFERENCE, CCNC, 2023,
  • [8] Communication-Efficient Federated Learning With Gradual Layer Freezing
    Malan, Erich
    Peluso, Valentino
    Calimera, Andrea
    Macii, Enrico
    IEEE EMBEDDED SYSTEMS LETTERS, 2023, 15 (01) : 25 - 28
  • [9] Communication-Efficient Federated Learning With Binary Neural Networks
    Yang, Yuzhi
    Zhang, Zhaoyang
    Yang, Qianqian
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (12) : 3836 - 3850
  • [10] Data Distribution-Aware Online Client Selection Algorithm for Federated Learning in Heterogeneous Networks
    Lee, Jaewook
    Ko, Haneul
    Seo, Sangwon
    Pack, Sangheon
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (01) : 1127 - 1136