FedNN: Federated learning on concept drift data using weight and adaptive group normalizations

被引:5
作者
Kang, Myeongkyun [1 ]
Kim, Soopil [1 ]
Jin, Kyong Hwan [2 ]
Adeli, Ehsan [3 ,4 ]
Pohl, Kilian M. [3 ]
Park, Sang Hyun [1 ,5 ]
机构
[1] Daegu Gyeongbuk Inst Sci & Technol DGIST, Dept Robot & Mechatron Engn, Daegu 42988, South Korea
[2] Korea Univ, Sch Elect Engn, Seoul, South Korea
[3] Stanford Univ, Dept Psychiat & Behav Sci, Stanford, CA 94305 USA
[4] Stanford Univ, Dept Comp Sci, Stanford, CA USA
[5] Daegu Gyeongbuk Inst Sci & Technol DGIST, AI Grad Sch, Daegu, South Korea
基金
新加坡国家研究基金会;
关键词
Federated learning; Concept drift; Weight normalization; Adaptive group normalization;
D O I
10.1016/j.patcog.2023.110230
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Federated Learning (FL) allows a global model to be trained without sharing private raw data. The major challenge in FL is client -wise data heterogeneity leading to different model convergence speed and accuracy. Despite the recent progress of FL, most methods verify their accuracy on prior probability shift (label distribution skew) dataset, while the concept drift problem (i.e., where each client has distinct styles of input while sharing the same labels) has not been explored. In real scenarios, concept drift is of paramount concern in FL since the client's data is collected under extremely different conditions making FL optimization more challenging. Significant differences in inputs among clients exacerbate the heterogeneity of clients' parameters compared to prior probability shift, ultimately resulting in failures for previous FL approaches. To address the challenge of concept drift, we use Weight Normalization (WN) and Adaptive Group Normalization (AGN) to alleviate conflicts during global model updates. WN re -parameterizes weights to have zero mean and unit variance while AGN adaptively selects the optimal mean and standard deviation for feature normalization based on the dataset. These two components significantly contribute to having consistent activations after global model updates reducing heterogeneity in concept drift data. Comprehensive experiments on seven datasets (with concept drift) demonstrate that our method outperforms five state-of-the-art FL methods and shows faster convergence speed compared to the previous FL methods.
引用
收藏
页数:11
相关论文
共 43 条
  • [1] Acar D.A.E., 2021, INT C LEARNING REPRE
  • [2] Ba Jimmy Lei, 2016, arXiv
  • [3] Coates Adam, 2011, NIPS WORKSHOP DEEP L
  • [4] Deecke L., 2019, INT C LEARNING REPRE
  • [5] Ganin Y, 2015, PR MACH LEARN RES, V37, P1180
  • [6] FedDC: Federated Learning with Non-IID Data via Local Drift Decoupling and Correction
    Gao, Liang
    Fu, Huazhu
    Li, Li
    Chen, Yingwen
    Xu, Ming
    Xu, Cheng-Zhong
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 10102 - 10111
  • [7] Representative Batch Normalization with Feature Calibration
    Gao, Shang-Hua
    Han, Qi
    Li, Duo
    Cheng, Ming-Ming
    Peng, Pai
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 8665 - 8675
  • [8] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [9] Hendrycks Dan, 2019, INT C LEARN REPR
  • [10] Centered Weight Normalization in Accelerating Training of Deep Neural Networks
    Huang, Lei
    Liu, Xianglong
    Liu, Yang
    Lang, Bo
    Tao, Dacheng
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2822 - 2830