How to Fine-Tune BERT for Text Classification?

被引:720
|
作者
Sun, Chi [1 ]
Qiu, Xipeng [1 ]
Xu, Yige [1 ]
Huang, Xuanjing [1 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai Key Lab Intelligent Informat Proc, 825 Zhangheng Rd, Shanghai, Peoples R China
来源
CHINESE COMPUTATIONAL LINGUISTICS, CCL 2019 | 2019年 / 11856卷
关键词
Transfer learning; BERT; Text classification;
D O I
10.1007/978-3-030-32381-3_16
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Language model pre-training has proven to be useful in learning universal language representations. As a state-of-the-art language model pre-training model, BERT (Bidirectional Encoder Representations from Transformers) has achieved amazing results in many language understanding tasks. In this paper, we conduct exhaustive experiments to investigate different fine-tuning methods of BERT on text classification task and provide a general solution for BERT fine-tuning. Finally, the proposed solution obtains new state-of-the-art results on eight widely-studied text classification datasets.
引用
收藏
页码:194 / 206
页数:13
相关论文
共 50 条
  • [41] A Multiscale Interactive Attention Short Text Classification Model Based on BERT
    Zhou, Lu
    Wang, Peng
    Zhang, Huijun
    Wu, Shengbo
    Zhang, Tao
    IEEE ACCESS, 2024, 12 : 160992 - 161001
  • [42] Three-Branch BERT-Based Text Classification Network for Gastroscopy Diagnosis Text
    Wang Z.
    Zheng X.
    Zhang J.
    Zhang M.
    International Journal of Crowd Science, 2024, 8 (01) : 56 - 63
  • [43] Research on Public Service Request Text Classification Based on BERT-BiLSTM-CNN Feature Fusion
    Xiong, Yunpeng
    Chen, Guolian
    Cao, Junkuo
    APPLIED SCIENCES-BASEL, 2024, 14 (14):
  • [44] Survey of BERT-Base Models for Scientific Text Classification: COVID-19 Case Study
    Khadhraoui, Mayara
    Bellaaj, Hatem
    Ben Ammar, Mehdi
    Hamam, Habib
    Jmaiel, Mohamed
    APPLIED SCIENCES-BASEL, 2022, 12 (06):
  • [45] Research on Internet Text Sentiment Classification Based on BERT and CNN-BiGRU
    Wei, Guoli
    2022 11TH INTERNATIONAL CONFERENCE ON COMMUNICATIONS, CIRCUITS AND SYSTEMS (ICCCAS 2022), 2022, : 285 - 289
  • [46] MII: A Novel Text Classification Model Combining Deep Active Learning with BERT
    Zhang, Anman
    Li, Bohan
    Wang, Wenhuan
    Wan, Shuo
    Chen, Weitong
    CMC-COMPUTERS MATERIALS & CONTINUA, 2020, 63 (03): : 1499 - 1514
  • [47] Financial causal sentence recognition based on BERT-CNN text classification
    Chang-Xuan Wan
    Bo Li
    The Journal of Supercomputing, 2022, 78 : 6503 - 6527
  • [48] Financial causal sentence recognition based on BERT-CNN text classification
    Wan, Chang-Xuan
    Li, Bo
    JOURNAL OF SUPERCOMPUTING, 2022, 78 (05) : 6503 - 6527
  • [49] Research on News Text Classification Based on BERT-BiLSTM-TextCNN-Attention
    Wang, Jia
    Li, Zongting
    Ma, Chenyang
    PROCEEDINGS OF 2024 3RD INTERNATIONAL CONFERENCE ON CYBER SECURITY, ARTIFICIAL INTELLIGENCE AND DIGITAL ECONOMY, CSAIDE 2024, 2024, : 295 - 298
  • [50] Text classification for distribution substation inspection based on BERT-TextRCNN model
    Lu, Jiangang
    Zhao, Ruifeng
    Yu, Zhiwen
    Dai, Yue
    Shu, Jiawei
    Yang, Ting
    FRONTIERS IN ENERGY RESEARCH, 2024, 12