A Multi-teacher Knowledge Distillation Framework for Distantly Supervised Relation Extraction with Flexible Temperature

被引:0
作者
Fei, Hongxiao [1 ]
Tan, Yangying [1 ]
Huang, Wenti [2 ]
Long, Jun [1 ,3 ]
Huang, Jincai [3 ]
Yang, Liu [1 ]
机构
[1] Cent South Univ, Sch Comp Sci & Engn, Changsha, Peoples R China
[2] Hunan Univ Sci & Technol, Sch Comp Sci & Engn, Xiangtan, Peoples R China
[3] Cent South Univ, Big Data Inst, Changsha, Peoples R China
来源
WEB AND BIG DATA, PT II, APWEB-WAIM 2023 | 2024年 / 14332卷
基金
中国国家自然科学基金;
关键词
Relation Extraction; Knowledge Distillation; Distantly Supervised; Multiple Teachers; Natural Language Processing;
D O I
10.1007/978-981-97-2390-4_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Distantly supervised relation extraction (DSRE) generates large-scale annotated data by aligning unstructured text with knowledge bases. However, automatic construction methods cause a substantial number of incorrect annotations, thereby introducing noise into the training process. Most sentence-level relation extraction methods rely on filters to remove noise instances, meanwhile, they ignore some useful information in negative instances. To effectively reduce noise interference, we propose a Multi-teacher Knowledge Distillation framework for Relation Extraction (MKDRE) to extract semantic relations from noisy data based on both global information and local information. MKDRE addresses two main problems: the deviation in knowledge propagation of a single teacher and the limitation of traditional distillation temperature on information utilization. Specifically, we utilize flexible temperature regulation (FTR) to adjust the temperature assigned to each training instance, so as to dynamically capture local relations between instances. Furthermore, we introduce information entropy of hidden layers to gain stable temperature calculations. Finally, we propose multi-view knowledge distillation (MVKD) to express global relations among teachers from various perspectives to gain more reliable knowledge. The experimental results on NYT19-1.0 and NYT19-2.0 datasets show that our proposed MKDRE significantly outperforms previous methods in sentence-level relation extraction.
引用
收藏
页码:103 / 116
页数:14
相关论文
共 36 条
  • [1] Chen Tao, 2021, P 59 ANN M ASS COMP, V1, P6191, DOI [10.18653/V1/2021.ACL, DOI 10.18653/V1/2021.ACL-LONG.483]
  • [2] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [3] Du JH, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P2216
  • [4] Feng J, 2018, AAAI CONF ARTIF INTE, P5779
  • [5] He ZQ, 2020, AAAI CONF ARTIF INTE, V34, P7927
  • [6] Hinton G., 2015, NIPS DEEP LEARN REPR
  • [7] Jia W, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P1399
  • [8] Lei, 2018, P 27 INT C COMP LING, P426
  • [9] Li DY, 2022, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), P2567
  • [10] MiDTD: A Simple and Effective Distillation Framework for Distantly Supervised Relation Extraction
    Li, Rui
    Yang, Cheng
    Li, Tingwei
    Su, Sen
    [J]. ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2022, 40 (04)