A Comparative Study on Pre-Trained Models Based on BERT

被引:0
作者
Zhang, Minghua [1 ]
机构
[1] Northeastern Univ, Khoury Coll Comp Sci, Beijing, Peoples R China
来源
2024 6TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING, ICNLP 2024 | 2024年
关键词
Self-Supervised Learning; PTM; NLP; BERT;
D O I
10.1109/ICNLP60986.2024.10692659
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The introduction of pre-trained models (PTMs) especially Bidirectional Encoder Representation from Transformer (BERT) [1] brought significant improvements in Natural Language Processing (NLP) tasks and demonstrated the power of transfer learning in large language models. The state-of-the-art performance of BERT on eleven NLP tasks inspired many researchers to focus on building variants based on BERT. This survey is going to collect and investigate the NLP-PTMs researches especially the ones motivated by BERT, concentrating on three main tasks: classifications of their research objects and research methods, and an experimental analysis. The collected papers are going to be classified based on different criteria for each task and provide detailed explanations of why certain research is classified into certain type. In the end, based on the investigation, a future direction for the development of PTMs in NLP is suggested.
引用
收藏
页码:326 / 330
页数:5
相关论文
共 31 条
  • [1] Chen J, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P4946
  • [2] Shao CC, 2019, Arxiv, DOI arXiv:1806.00920
  • [3] Choi E, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P2174
  • [4] ELECTRA: PRE-TRAINING TEXT ENCODERS AS DISCRIMINATORS RATHER THAN GENERATORS
    Clark, Kevin
    Luong, Minh-Thang
    Le, Quoc V.
    Manning, Christopher D.
    [J]. INFORMATION SYSTEMS RESEARCH, 2020,
  • [5] Cui YM, 2021, Arxiv, DOI [arXiv:1906.08101, 10.48550/arXiv.1906.08101]
  • [6] Cui YM, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P5883
  • [7] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [8] CJRC: A Reliable Human-Annotated Benchmark DataSet for Chinese Judicial Reading Comprehension
    Duan, Xingyi
    Wang, Baoxin
    Wang, Ziyue
    Ma, Wentao
    Cui, Yiming
    Wu, Dayong
    Wang, Shijin
    Liu, Ting
    Huo, Tianxiang
    Hu, Zhen
    Wang, Heng
    Liu, Zhiyuan
    [J]. CHINESE COMPUTATIONAL LINGUISTICS, CCL 2019, 2019, 11856 : 439 - 451
  • [9] Hu H, 2020, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020
  • [10] SpanBERT: Improving Pre-training by Representing and Predicting Spans
    Joshi, Mandar
    Chen, Danqi
    Liu, Yinhan
    Weld, Daniel S.
    Zettlemoyer, Luke
    Levy, Omer
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2020, 8 : 64 - 77