Federated Split BERT for Heterogeneous Text Classification

被引:3
|
作者
Li, Zhengyang [1 ]
Si, Shijing [1 ]
Wang, Jianzong [1 ]
Xiao, Jing [1 ]
机构
[1] Ping An Technol Shenzhen Co Ltd, Shenzhen, Peoples R China
来源
2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN) | 2022年
关键词
Federated Learning; BERT; Data Heterogeneity; Quantization; Text Classification;
D O I
10.1109/IJCNN55064.2022.9892845
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained BERT models have achieved impressive performance in many natural language processing (NLP) tasks. However, in many real-world situations, textual data are usually decentralized over many clients and unable to be uploaded to a central server due to privacy protection and regulations. Federated learning (FL) enables multiple clients collaboratively to train a global model while keeping the local data privacy. A few researches have investigated BERT in federated learning setting, but the problem of performance loss caused by heterogeneous (e.g., non-IID) data over clients remain under-explored. To address this issue, we propose a framework, FedSplitBERT, which handles heterogeneous data and decreases the communication cost by splitting the BERT encoder layers into local part and global part. The local part parameters are trained by the local client only while the global part parameters are trained by aggregating gradients of multiple clients. Due to the sheer size of BERT, we explore a quantization method to further reduce the communication cost with minimal performance loss. Our framework is ready-to-use and compatible to many existing federated learning algorithms, including FedAvg, FedProx and FedAdam. Our experiments verify the effectiveness of the proposed framework, which outperforms baseline methods by a significant margin, while FedSplitBERT with quantization can reduce the communication cost by 11.9x.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Federated Freeze BERT for text classification
    Omar Galal
    Ahmed H. Abdel-Gawad
    Mona Farouk
    Journal of Big Data, 11
  • [2] Federated Freeze BERT for text classification
    Galal, Omar
    Abdel-Gawad, Ahmed H.
    Farouk, Mona
    JOURNAL OF BIG DATA, 2024, 11 (01)
  • [3] Turkish Medical Text Classification Using BERT
    Celikten, Azer
    Bulut, Hasan
    29TH IEEE CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS (SIU 2021), 2021,
  • [4] Rethinking of BERT sentence embedding for text classification
    Omar Galal
    Ahmed H. Abdel-Gawad
    Mona Farouk
    Neural Computing and Applications, 2024, 36 (32) : 20245 - 20258
  • [5] An Automated Text Document Classification Framework using BERT
    Shah, Momna Ali
    Iqbal, Muhammad Javed
    Noreen, Neelum
    Ahmed, Iftikhar
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2023, 14 (03) : 279 - 285
  • [6] How to Fine-Tune BERT for Text Classification?
    Sun, Chi
    Qiu, Xipeng
    Xu, Yige
    Huang, Xuanjing
    CHINESE COMPUTATIONAL LINGUISTICS, CCL 2019, 2019, 11856 : 194 - 206
  • [7] WHEN BERT MEETS QUANTUM TEMPORAL CONVOLUTION LEARNING FOR TEXT CLASSIFICATION IN HETEROGENEOUS COMPUTING
    Yang, Chao-Han Huck
    Qi, Jun
    Chen, Samuel Yen-Chi
    Tsao, Yu
    Chen, Pin-Yu
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8602 - 8606
  • [8] Effective text classification using BERT, MTM LSTM, and DT
    Jamshidi, Saman
    Mohammadi, Mahin
    Bagheri, Saeed
    Najafabadi, Hamid Esmaeili
    Rezvanian, Alireza
    Gheisari, Mehdi
    Ghaderzadeh, Mustafa
    Shahabi, Amir Shahab
    Wu, Zongda
    DATA & KNOWLEDGE ENGINEERING, 2024, 151
  • [9] Bert-Enhanced Text Graph Neural Network for Classification
    Yang, Yiping
    Cui, Xiaohui
    ENTROPY, 2021, 23 (11)
  • [10] The Automatic Text Classification Method Based on BERT and Feature Union
    Li, Wenting
    Gao, Shangbing
    Zhou, Hong
    Huang, Zihe
    Zhang, Kewen
    Li, Wei
    2019 IEEE 25TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2019, : 774 - 777