BERT-Based Transfer-Learning Approach for Nested Named-Entity Recognition Using Joint Labeling

被引:25
|
作者
Agrawal, Ankit [1 ]
Tripathi, Sarsij [2 ]
Vardhan, Manu [1 ]
Sihag, Vikas [3 ]
Choudhary, Gaurav [4 ]
Dragoni, Nicola [4 ]
机构
[1] Natl Inst Technol Raipur, Dept Comp Sci & Engn, Raipur 492010, Chhattisgarh, India
[2] Motilal Nehru Natl Inst Technol Allahabad, Dept Comp Sci & Engn, Prayagraj 211004, Uttar Pradesh, India
[3] Sardar Patel Univ Police, Dept Cyber Secur, Secur & Criminal Justice, Jodhpur 342037, Rajasthan, India
[4] Tech Univ Denmark DTU, Dept Appl Math & Comp Sci, DTU Comp, DK-2800 Lyngby, Denmark
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 03期
关键词
named-entity recognition; transfer learning; BERT model; conditional random field; pre-trained model; fine-tuning;
D O I
10.3390/app12030976
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Named-entity recognition (NER) is one of the primary components in various natural language processing tasks such as relation extraction, information retrieval, question answering, etc. The majority of the research work deals with flat entities. However, it was observed that the entities were often embedded within other entities. Most of the current state-of-the-art models deal with the problem of embedded/nested entity recognition with very complex neural network architectures. In this research work, we proposed to solve the problem of nested named-entity recognition using the transfer-learning approach. For this purpose, different variants of fine-tuned, pretrained, BERT-based language models were used for the problem using the joint-labeling modeling technique. Two nested named-entity-recognition datasets, i.e., GENIA and GermEval 2014, were used for the experiment, with four and two levels of annotation, respectively. Also, the experiments were performed on the JNLPBA dataset, which has flat annotation. The performance of the above models was measured using F1-score metrics, commonly used as the standard metrics to evaluate the performance of named-entity-recognition models. In addition, the performance of the proposed approach was compared with the conditional random field and the Bi-LSTM-CRF model. It was found that the fine-tuned, pretrained, BERT-based models outperformed the other models significantly without requiring any external resources or feature extraction. The results of the proposed models were compared with various other existing approaches. The best-performing BERT-based model achieved F1-scores of 74.38, 85.29, and 80.68 for the GENIA, GermEval 2014, and JNLPBA datasets, respectively. It was found that the transfer learning (i.e., pretrained BERT models after fine-tuning) based approach for the nested named-entity-recognition task could perform well and is a more generalized approach in comparison to many of the existing approaches.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] ATBBC: Named entity recognition in emergency domains based on joint BERT-BILSTM-CRF adversarial training
    Cai, Buqing
    Tian, Shengwei
    Yu, Long
    Long, Jun
    Zhou, Tiejun
    Wang, Bo
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2024, 46 (02) : 4063 - 4076
  • [32] Negation-based transfer learning for improving biomedical Named Entity Recognition and Relation Extraction
    Fabregat, Hermenegildo
    Duque, Andres
    Martinez-Romo, Juan
    Araujo, Lourdes
    JOURNAL OF BIOMEDICAL INFORMATICS, 2023, 138
  • [33] Arabic Location Named Entity Recognition for Tweets using a Deep Learning Approach
    Alzaidi, Bedour Swayelh
    Abushark, Yoosef
    Khan, Asif Irshad
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (12) : 76 - 83
  • [34] Transformer-based approach for joint handwriting and named entity recognition in historical document
    Rouhou, Ahmed Cheikh
    Dhiaf, Marwa
    Kessentini, Yousri
    Ben Salem, Sinda
    PATTERN RECOGNITION LETTERS, 2022, 155 : 128 - 134
  • [35] Few-shot Learning for Named Entity Recognition Based on BERT and Two-level Model Fusion
    Gong, Yuan
    Mao, Lu
    Li, Changliang
    DATA INTELLIGENCE, 2021, 3 (04) : 568 - 577
  • [36] Named entity recognition of local adverse drug reactions in Xinjiang based on transfer learning
    Kang, Keming
    Tian, Shengwei
    Yu, Long
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 40 (05) : 8899 - 8914
  • [37] A proposal to identify stakeholders from news for the institutional relationship management activities of an institution based on Named Entity Recognition using BERT
    Messias da Silva, Eric Hans
    Laterza, Joao
    Pereira da Silva, Marcos Paulo
    Ladeira, Marcelo
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 1569 - 1575
  • [38] Thai Named-Entity Recognition Using Class-based Language Modeling on Multiple-sized Subword Units
    Saykhum, Kwanchiva
    Boonpiam, Vataya
    Thatphithakkul, Nattanun
    Wutiwiwatchai, Chai
    Natthee, Cholwich
    INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 1586 - +
  • [39] Named entity recognition on bio-medical literature documents using hybrid based approach
    Ramachandran, R.
    Arutchelvan, K.
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021,
  • [40] Active Learning-Based Approach for Named Entity Recognition on Short Text Streams
    Cuong Van Tran
    Tuong Tri Nguyen
    Dinh Tuyen Hoang
    Hwang, Dosam
    Ngoc Thanh Nguyen
    MULTIMEDIA AND NETWORK INFORMATION SYSTEMS, MISSI 2016, 2017, 506 : 321 - 330