Bidirectional Decoupled Distillation for Heterogeneous Federated Learning

被引:0
作者
Song, Wenshuai [1 ]
Yan, Mengwei [1 ]
Li, Xinze [1 ]
Han, Longfei [1 ]
机构
[1] Beijing Technol & Business Univ, Sch Comp & Artificial Intelligence, Beijing 100048, Peoples R China
基金
中国国家自然科学基金;
关键词
information theory; personalized federated learning; knowledge distillation; PRIVACY;
D O I
10.3390/e26090762
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Federated learning enables multiple devices to collaboratively train a high-performance model on the central server while keeping their data on the devices themselves. However, due to the significant variability in data distribution across devices, the aggregated global model's optimization direction may differ from that of the local models, making the clients lose their personality. To address this challenge, we propose a Bidirectional Decoupled Distillation For Heterogeneous Federated Learning (BDD-HFL) approach, which incorporates an additional private model within each local client. This design enables mutual knowledge exchange between the private and local models in a bidirectional manner. Specifically, previous one-way federated distillation methods mainly focused on learning features from the target class, which limits their ability to distill features from non-target classes and hinders the convergence of local models. To solve this limitation, we decompose the network output into target and non-target class logits and distill them separately using a joint optimization of cross-entropy and decoupled relative-entropy loss. We evaluate the effectiveness of BDD-HFL through extensive experiments on three benchmarks under IID, Non-IID, and unbalanced data distribution scenarios. Our results show that BDD-HFL outperforms state-of-the-art federated distillation methods across five baselines, achieving at most 3% improvement in average classification accuracy on the CIFAR-10, CIFAR-100, and MNIST datasets. The experiments demonstrate the superiority and generalization capability of BDD-HFL in addressing personalization challenges in federated learning.
引用
收藏
页数:14
相关论文
共 45 条
[1]   Ensemble unsupervised autoencoders and Gaussian mixture model for cyberattack detection [J].
An, Peng ;
Wang, Zhiyuan ;
Zhang, Chunjiong .
INFORMATION PROCESSING & MANAGEMENT, 2022, 59 (02)
[2]   Data-Free Learning of Student Networks [J].
Chen, Hanting ;
Wang, Yunhe ;
Xu, Chang ;
Yang, Zhaohui ;
Liu, Chuanjian ;
Shi, Boxin ;
Xu, Chunjing ;
Xu, Chao ;
Tian, Qi .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3513-3521
[3]   Convergence Time Optimization for Federated Learning Over Wireless Networks [J].
Chen, Mingzhe ;
Poor, H. Vincent ;
Saad, Walid ;
Cui, Shuguang .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (04) :2457-2471
[4]  
Deng Y., 2021, P INT C LEARN REPR V
[5]  
Diao E., 2021, P INT C LEARN REPR V
[6]  
Dinh CT, 2020, ADV NEUR IN, V33
[7]  
Durmus A.E., 2021, P INT C LEARN REPR V
[8]  
Faisal F., 2023, P 2023 IEEE 10 INT C, P1
[9]  
Fallah A, 2020, ADV NEUR IN, V33
[10]   FedDC: Federated Learning with Non-IID Data via Local Drift Decoupling and Correction [J].
Gao, Liang ;
Fu, Huazhu ;
Li, Li ;
Chen, Yingwen ;
Xu, Ming ;
Xu, Cheng-Zhong .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :10102-10111