Federated Split BERT for Heterogeneous Text Classification

被引:3
|
作者
Li, Zhengyang [1 ]
Si, Shijing [1 ]
Wang, Jianzong [1 ]
Xiao, Jing [1 ]
机构
[1] Ping An Technol Shenzhen Co Ltd, Shenzhen, Peoples R China
来源
2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN) | 2022年
关键词
Federated Learning; BERT; Data Heterogeneity; Quantization; Text Classification;
D O I
10.1109/IJCNN55064.2022.9892845
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained BERT models have achieved impressive performance in many natural language processing (NLP) tasks. However, in many real-world situations, textual data are usually decentralized over many clients and unable to be uploaded to a central server due to privacy protection and regulations. Federated learning (FL) enables multiple clients collaboratively to train a global model while keeping the local data privacy. A few researches have investigated BERT in federated learning setting, but the problem of performance loss caused by heterogeneous (e.g., non-IID) data over clients remain under-explored. To address this issue, we propose a framework, FedSplitBERT, which handles heterogeneous data and decreases the communication cost by splitting the BERT encoder layers into local part and global part. The local part parameters are trained by the local client only while the global part parameters are trained by aggregating gradients of multiple clients. Due to the sheer size of BERT, we explore a quantization method to further reduce the communication cost with minimal performance loss. Our framework is ready-to-use and compatible to many existing federated learning algorithms, including FedAvg, FedProx and FedAdam. Our experiments verify the effectiveness of the proposed framework, which outperforms baseline methods by a significant margin, while FedSplitBERT with quantization can reduce the communication cost by 11.9x.
引用
收藏
页数:8
相关论文
共 50 条
  • [41] MII: A Novel Text Classification Model Combining Deep Active Learning with BERT
    Zhang, Anman
    Li, Bohan
    Wang, Wenhuan
    Wan, Shuo
    Chen, Weitong
    CMC-COMPUTERS MATERIALS & CONTINUA, 2020, 63 (03): : 1499 - 1514
  • [42] Financial causal sentence recognition based on BERT-CNN text classification
    Chang-Xuan Wan
    Bo Li
    The Journal of Supercomputing, 2022, 78 : 6503 - 6527
  • [43] Financial causal sentence recognition based on BERT-CNN text classification
    Wan, Chang-Xuan
    Li, Bo
    JOURNAL OF SUPERCOMPUTING, 2022, 78 (05) : 6503 - 6527
  • [44] Research on News Text Classification Based on BERT-BiLSTM-TextCNN-Attention
    Wang, Jia
    Li, Zongting
    Ma, Chenyang
    PROCEEDINGS OF 2024 3RD INTERNATIONAL CONFERENCE ON CYBER SECURITY, ARTIFICIAL INTELLIGENCE AND DIGITAL ECONOMY, CSAIDE 2024, 2024, : 295 - 298
  • [45] Text classification for distribution substation inspection based on BERT-TextRCNN model
    Lu, Jiangang
    Zhao, Ruifeng
    Yu, Zhiwen
    Dai, Yue
    Shu, Jiawei
    Yang, Ting
    FRONTIERS IN ENERGY RESEARCH, 2024, 12
  • [46] Improving BERT-Based Text Classification With Auxiliary Sentence and Domain Knowledge
    Yu, Shanshan
    Su, Jindian
    Luo, Da
    IEEE ACCESS, 2019, 7 : 176600 - 176612
  • [47] MII: A novel text classification model combining deep active learning with BERT
    Zhang A.
    Li B.
    Wang W.
    Wan S.
    Chen W.
    Computers, Materials and Continua, 2020, 63 (03): : 1499 - 1514
  • [48] Text Classification Model Based on BERT-Capsule with Integrated Deep Learning
    Tian, Yuwei
    Zhang, Zhi
    PROCEEDINGS OF THE 2021 IEEE 16TH CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA 2021), 2021, : 106 - 111
  • [49] Federated or Split? A Performance and Privacy Analysis of Hybrid Split and Federated Learning Architectures
    Turina, Valeria
    Zhang, Zongshun
    Esposito, Flavio
    Matta, Ibrahim
    2021 IEEE 14TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING (CLOUD 2021), 2021, : 250 - 260
  • [50] Deeply integrating unsupervised semantics and syntax into heterogeneous graphs for inductive text classification
    Yue Gao
    Xiangling Fu
    Xien Liu
    Ji Wu
    Complex & Intelligent Systems, 2024, 10 : 1565 - 1579