BERT-Based Transfer-Learning Approach for Nested Named-Entity Recognition Using Joint Labeling

被引:25
|
作者
Agrawal, Ankit [1 ]
Tripathi, Sarsij [2 ]
Vardhan, Manu [1 ]
Sihag, Vikas [3 ]
Choudhary, Gaurav [4 ]
Dragoni, Nicola [4 ]
机构
[1] Natl Inst Technol Raipur, Dept Comp Sci & Engn, Raipur 492010, Chhattisgarh, India
[2] Motilal Nehru Natl Inst Technol Allahabad, Dept Comp Sci & Engn, Prayagraj 211004, Uttar Pradesh, India
[3] Sardar Patel Univ Police, Dept Cyber Secur, Secur & Criminal Justice, Jodhpur 342037, Rajasthan, India
[4] Tech Univ Denmark DTU, Dept Appl Math & Comp Sci, DTU Comp, DK-2800 Lyngby, Denmark
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 03期
关键词
named-entity recognition; transfer learning; BERT model; conditional random field; pre-trained model; fine-tuning;
D O I
10.3390/app12030976
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Named-entity recognition (NER) is one of the primary components in various natural language processing tasks such as relation extraction, information retrieval, question answering, etc. The majority of the research work deals with flat entities. However, it was observed that the entities were often embedded within other entities. Most of the current state-of-the-art models deal with the problem of embedded/nested entity recognition with very complex neural network architectures. In this research work, we proposed to solve the problem of nested named-entity recognition using the transfer-learning approach. For this purpose, different variants of fine-tuned, pretrained, BERT-based language models were used for the problem using the joint-labeling modeling technique. Two nested named-entity-recognition datasets, i.e., GENIA and GermEval 2014, were used for the experiment, with four and two levels of annotation, respectively. Also, the experiments were performed on the JNLPBA dataset, which has flat annotation. The performance of the above models was measured using F1-score metrics, commonly used as the standard metrics to evaluate the performance of named-entity-recognition models. In addition, the performance of the proposed approach was compared with the conditional random field and the Bi-LSTM-CRF model. It was found that the fine-tuned, pretrained, BERT-based models outperformed the other models significantly without requiring any external resources or feature extraction. The results of the proposed models were compared with various other existing approaches. The best-performing BERT-based model achieved F1-scores of 74.38, 85.29, and 80.68 for the GENIA, GermEval 2014, and JNLPBA datasets, respectively. It was found that the transfer learning (i.e., pretrained BERT models after fine-tuning) based approach for the nested named-entity-recognition task could perform well and is a more generalized approach in comparison to many of the existing approaches.
引用
收藏
页数:20
相关论文
共 50 条
  • [41] Named entity recognition using transfer learning and small human- and meta-pseudo-labeled datasets
    Bae, Kyoungman
    Lim, Joon-Ho
    ETRI JOURNAL, 2024, 46 (01) : 59 - 70
  • [42] Named Entity Recognition for Long COVID Biomedical Literature by Using Bert-BiLSTM-IDCNN-ATT-CRF Approach
    Han, Zongwang
    Lin, Shaofu
    Huang, Zhisheng
    Guo, Chaohui
    PROCEEDINGS OF 2023 4TH INTERNATIONAL SYMPOSIUM ON ARTIFICIAL INTELLIGENCE FOR MEDICINE SCIENCE, ISAIMS 2023, 2023, : 1200 - 1205
  • [43] Transfer Learning for Named Entity Recognition in Setswana Language Using CNN-BiLSTM Model
    Chabalala, Shumile
    Ojo, Sunday O.
    Owolawi, Pius A.
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2025, 16 (02) : 472 - 481
  • [44] Active learning approach using a modified least confidence sampling strategy for named entity recognition
    Agrawal, Ankit
    Tripathi, Sarsij
    Vardhan, Manu
    PROGRESS IN ARTIFICIAL INTELLIGENCE, 2021, 10 (02) : 113 - 128
  • [45] Active learning approach using a modified least confidence sampling strategy for named entity recognition
    Ankit Agrawal
    Sarsij Tripathi
    Manu Vardhan
    Progress in Artificial Intelligence, 2021, 10 : 113 - 128
  • [46] A Feature Based Simple Machine Learning Approach with Word Embeddings to Named Entity Recognition on Tweets
    Taspinar, Mete
    Ganiz, Murat Can
    Acarman, Tankut
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS, NLDB 2017, 2017, 10260 : 254 - 259
  • [47] Supporting Deep Learning-Based Named Entity Recognition Using Cloud Resource Management
    Hartmann, Benedict
    Tamla, Philippe
    Hemmje, Matthias
    HCI INTERNATIONAL 2023 LATE BREAKING PAPERS, HCII 2023, PT VI, 2023, 14059 : 84 - 100
  • [48] Application of named entity recognition on tweets during earthquake disaster: a deep learning-based approach
    Eliguzel, Nazmiye
    Cetinkaya, Cihan
    Dereli, Turkay
    SOFT COMPUTING, 2022, 26 (01) : 395 - 421
  • [49] Application of named entity recognition on tweets during earthquake disaster: a deep learning-based approach
    Nazmiye Eligüzel
    Cihan Çetinkaya
    Türkay Dereli
    Soft Computing, 2022, 26 : 395 - 421
  • [50] External Knowledge-Based Weakly Supervised Learning Approach on Chinese Clinical Named Entity Recognition
    Duan, Yeheng
    Ma, Long-Long
    Han, Xianpei
    Sun, Le
    Dong, Bin
    Jiang, Shanshan
    SEMANTIC TECHNOLOGY, JIST 2019: PROCEEDINGS, 2020, 12032 : 336 - 352