Federated Split BERT for Heterogeneous Text Classification

被引:3
|
作者
Li, Zhengyang [1 ]
Si, Shijing [1 ]
Wang, Jianzong [1 ]
Xiao, Jing [1 ]
机构
[1] Ping An Technol Shenzhen Co Ltd, Shenzhen, Peoples R China
来源
2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN) | 2022年
关键词
Federated Learning; BERT; Data Heterogeneity; Quantization; Text Classification;
D O I
10.1109/IJCNN55064.2022.9892845
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained BERT models have achieved impressive performance in many natural language processing (NLP) tasks. However, in many real-world situations, textual data are usually decentralized over many clients and unable to be uploaded to a central server due to privacy protection and regulations. Federated learning (FL) enables multiple clients collaboratively to train a global model while keeping the local data privacy. A few researches have investigated BERT in federated learning setting, but the problem of performance loss caused by heterogeneous (e.g., non-IID) data over clients remain under-explored. To address this issue, we propose a framework, FedSplitBERT, which handles heterogeneous data and decreases the communication cost by splitting the BERT encoder layers into local part and global part. The local part parameters are trained by the local client only while the global part parameters are trained by aggregating gradients of multiple clients. Due to the sheer size of BERT, we explore a quantization method to further reduce the communication cost with minimal performance loss. Our framework is ready-to-use and compatible to many existing federated learning algorithms, including FedAvg, FedProx and FedAdam. Our experiments verify the effectiveness of the proposed framework, which outperforms baseline methods by a significant margin, while FedSplitBERT with quantization can reduce the communication cost by 11.9x.
引用
收藏
页数:8
相关论文
共 50 条
  • [21] BVMHA: Text classification model with variable multihead hybrid attention based on BERT
    Peng, Bo
    Zhang, Tao
    Han, Kundong
    Zhang, Zhe
    Ma, Yuquan
    Ma, Mengnan
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2024, 46 (01) : 1443 - 1454
  • [22] Understanding the Impact of Label Skewness and Optimization on Federated Learning for Text Classification
    Francis, Sumam
    Uma, Kanimozhi
    Moens, Marie-Francine
    COMPANION OF THE WORLD WIDE WEB CONFERENCE, WWW 2023, 2023, : 1161 - 1166
  • [23] FF-BERT: A BERT-based ensemble for automated classification of web-based text on flash flood events
    Wilkho, Rohan Singh
    Chang, Shi
    Gharaibeh, Nasir G.
    ADVANCED ENGINEERING INFORMATICS, 2024, 59
  • [24] Cross-Domain Text Classification Based on BERT Model
    Zhang, Kuan
    Hei, Xinhong
    Fei, Rong
    Guo, Yufan
    Jiao, Rui
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS: DASFAA 2021 INTERNATIONAL WORKSHOPS, 2021, 12680 : 197 - 208
  • [25] Text Classification Research Based on Bert Model and Bayesian Network
    Liu, Songsong
    Tao, Haijun
    Feng, Shiling
    2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 5842 - 5846
  • [26] Analyzing the Performance of BERT for the Sentiment Classification Task in Bengali Text
    Banshal, Sumit Kumar
    Uddin, Ashraf
    Piryani, Rajesh
    ADVANCED NETWORK TECHNOLOGIES AND INTELLIGENT COMPUTING, ANTIC 2023, PT III, 2024, 2092 : 273 - 285
  • [27] Text Classification by CEFR Levels Using Machine Learning Methods and the BERT Language Model
    Lagutina, N. S.
    Lagutina, K. V.
    Brederman, A. M.
    Kasatkina, N. N.
    AUTOMATIC CONTROL AND COMPUTER SCIENCES, 2024, 58 (07) : 869 - 878
  • [28] Text classification problems via BERT embedding method and graph convolutional neural network
    Loc Tran
    Lam Pham
    Tuan Tran
    An Mai
    2021 INTERNATIONAL CONFERENCE ON ADVANCED TECHNOLOGIES FOR COMMUNICATIONS (ATC 2021), 2021, : 260 - 264
  • [29] A gating context-aware text classification model with BERT and graph convolutional networks
    Gao, Weiqi
    Huang, Hao
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 40 (03) : 4331 - 4343
  • [30] Emotion Classification of Text Based on BERT and Broad Learning System
    Peng, Sancheng
    Zeng, Rong
    Liu, Hongzhan
    Chen, Guanghao
    Wu, Ruihuan
    Yang, Aimin
    Yu, Shui
    WEB AND BIG DATA, APWEB-WAIM 2021, PT I, 2021, 12858 : 382 - 396