Federated Class-Incremental Learning via Weighted Aggregation and Distillation

被引:0
作者
Wu, Feng [1 ]
Ziying Tan, Alysa [2 ]
Feng, Siwei [1 ]
Yu, Han
Deng, Tao [1 ,3 ]
Zhao, Libang [1 ]
Chen, Yuanlu [1 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou 215000, Peoples R China
[2] Nanyang Technol Univ, Coll Comp & Data Sci, Nanyang Ave, Singapore 639798, Singapore
[3] Southwest Jiaotong Univ, Key Lab Photon Elect Integrat & Commun Sensing Con, Minist Educ, Chengdu 610031, Peoples R China
基金
中国国家自然科学基金; 新加坡国家研究基金会;
关键词
Data models; Training; Noise measurement; Federated learning; Data privacy; Computational modeling; Servers; Accuracy; Synthetic data; Soft sensors; Catastrophic forgetting; data heterogeneity; federated class-incremental learning (FCIL); knowledge distillation; weighted aggregation;
D O I
10.1109/JIOT.2025.3553901
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated Class-Incremental Learning (FCIL) aims to design privacy-preserving collaborative training methods to continuously learn new classes from distributed datasets. In these scenarios, federated clients face the challenge of encountering new classes while being constrained by limited memory capacity, which can lead to catastrophic forgetting in the resulting global model. Existing FCIL approaches tend to overlook the challenges posed by the heterogeneity of dataset label distribution among clients, thereby constraining the generalization capacity of the global model they learn. Some of these methods also suffer from excessive computational burdens when addressing catastrophic forgetting problems. Furthermore, certain approaches are constrained to handling only straightforward data, posing significant difficulties in managing complex datasets and tackling more intricate scenarios. In this article, we propose the Weighted Aggregation and Distillation-based FCIL (WAD-FCIL) method to address these limitations. To address data heterogeneity arising from class imbalance, we first introduce a task-aware client clustering method to identify clients with extreme class deviations before global model aggregation to eliminate potential impact on the global model. Then, we propose a multisampling weighted aggregation approach during the global FL model update that integrates knowledge from different clients and dynamically adjusts the weight of each client model to facilitate model update. To mitigate catastrophic forgetting, we propose a multimodel distillation strategy that involves selecting multiple teacher models for knowledge distillation. Extensive experiments comparing WAD-FCIL with ten state-of-the-art methods demonstrate that it significantly outperforms the baselines by 0.8%-3.2% in terms of average test accuracy on three representative benchmark datasets. The code of this work is available at https://github.com/wufeng10010/WAD-FCIL.
引用
收藏
页码:22489 / 22503
页数:15
相关论文
共 71 条
[1]   Conditional Channel Gated Networks for Task-Aware Continual Learning [J].
Abati, Davide ;
Tomczak, Jakub ;
Blankevoort, Tijmen ;
Calderara, Simone ;
Cucchiara, Rita ;
Bejnordi, Babak Ehteshami .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :3930-3939
[2]   SS-IL: Separated Softmax for Incremental Learning [J].
Ahn, Hongjoon ;
Kwak, Jihwan ;
Lim, Subin ;
Bang, Hyeonsu ;
Kim, Hyojun ;
Moon, Taesup .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :824-833
[3]   Variational Information Distillation for Knowledge Transfer [J].
Ahn, Sungsoo ;
Hu, Shell Xu ;
Damianou, Andreas ;
Lawrence, Neil D. ;
Dai, Zhenwen .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9155-9163
[4]  
Aledhari M, 2020, IEEE ACCESS, V8, P140699, DOI [10.1109/ACCESS.2020.3013541, 10.1109/access.2020.3013541]
[5]  
Allen-Zhu Z, 2021, Arxiv, DOI [arXiv:2012.09816, DOI 10.48550/ARXIV.2012.09816]
[6]   Knowledge distillation: A good teacher is patient and consistent [J].
Beyer, Lucas ;
Zhai, Xiaohua ;
Royer, Amelie ;
Markeeva, Larisa ;
Anil, Rohan ;
Kolesnikov, Alexander .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :10915-10924
[7]   End-to-End Incremental Learning [J].
Castro, Francisco M. ;
Marin-Jimenez, Manuel J. ;
Guil, Nicolas ;
Schmid, Cordelia ;
Alahari, Karteek .
COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 :241-257
[8]  
Chen HJ, 2020, ADV NEUR IN, V33
[9]   Communication-Efficient Federated Deep Learning With Layerwise Asynchronous Model Update and Temporally Weighted Aggregation [J].
Chen, Yang ;
Sun, Xiaoyan ;
Jin, Yaochu .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (10) :4229-4238
[10]   Learning without Memorizing [J].
Dhar, Prithviraj ;
Singh, Rajat Vikram ;
Peng, Kuan-Chuan ;
Wu, Ziyan ;
Chellappa, Rama .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :5133-5141