ClassTer: Mobile Shift-Robust Personalized Federated Learning via Class-Wise Clustering

被引:0
作者
Li, Xiaochen [1 ]
Liu, Sicong [1 ]
Zhou, Zimu [2 ]
Xu, Yuan [1 ]
Guo, Bin [1 ]
Yu, Zhiwen [2 ,3 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci, Xian 710072, Peoples R China
[2] City Univ Hong Kong, Dept Data Sci, Hong Kong 999077, Peoples R China
[3] Harbin Engn Univ, Sch Comp Sci, Harbin 150001, Peoples R China
基金
中国国家自然科学基金;
关键词
Data models; Training; Adaptation models; Mobile handsets; Federated learning; Computational modeling; Convergence; Accuracy; Servers; Mobile applications; Asynchronous mobile devices; personalized federated learning; shift-robust;
D O I
10.1109/TMC.2024.3487294
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The rise of mobile devices with abundant sensor data and computing power has driven the trend of federated learning (FL) on them. Personalized FL (PFL) aims to train tailored models for each device, addressing data heterogeneity from diverse user behaviors and preferences. However, due to dynamic mobile environments, PFL faces challenges in test-time data shifts, i.e., variations between training and testing. While this issue is well studied in generic deep learning through model generalization or adaptation, this issue remains less explored in PFL, where models often overfit local data. To address this, we introduce ${\sf ClassTer}$ClassTer, a shift-robust PFL framework. We observe that class-wise clustering of clients in cluster-based PFL (CFL) can avoid class-specific biases by decoupling the training of classes. Thus, we propose a paradigm shift from traditional client-wise clustering to class-wise clustering, which allows effective aggregation of cluster models into a generalized one via knowledge distillation. Additionally, we extend ClassTer to asynchronous mobile clients to optimize wall clock time by leveraging critical learning periods and both intra- and inter-device scheduling. Experiments show that compared to status quo approaches, ${\sf ClassTer}$ClassTer achieves a reduction of up to 91% in convergence time, and an improvement of up to 50.45% in accuracy.
引用
收藏
页码:2014 / 2028
页数:15
相关论文
共 80 条
  • [1] [Anonymous], 2023, YouTube
  • [2] FedPacket: A Federated Learning Approach to Mobile Packet Classification
    Bakopoulou, Evita
    Tillman, Balint
    Markopoulou, Athina
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2022, 21 (10) : 3609 - 3628
  • [3] Canales L., 2014, P WORKSH NAT LANG PR, P37
  • [4] Contrastive Test-Time Adaptation
    Chen, Dian
    Wang, Dequan
    Darrell, Trevor
    Ibrahimi, Sayna
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 295 - 305
  • [5] nnPerf: Demystifying DNN Runtime Inference Latency on Mobile Platforms
    Chu, Haolin
    Zheng, Xiaolong
    Liu, Liang
    Ma, Huadong
    [J]. PROCEEDINGS OF THE 21ST ACM CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS, SENSYS 2023, 2023, : 125 - 137
  • [6] Deng YY, 2020, Arxiv, DOI arXiv:2003.13461
  • [7] Fallah A, 2020, ADV NEUR IN, V33
  • [8] Fang C, 2024, Arxiv, DOI arXiv:2410.08256
  • [9] Howard AG, 2017, Arxiv, DOI [arXiv:1704.04861, DOI 10.48550/ARXIV.1704.04861]
  • [10] An Efficient Framework for Clustered Federated Learning
    Ghosh, Avishek
    Chung, Jichan
    Yin, Dong
    Ramchandran, Kannan
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2022, 68 (12) : 8076 - 8091