Denoising Distantly Supervised Named Entity Recognition via a Hypergeometric Probabilistic Model

被引:0
作者
Zhang, Wenkai [1 ,3 ]
Lin, Hongyu [1 ]
Han, Xianpei [1 ,2 ]
Sun, Le [1 ,2 ]
Liu, Huidan [1 ]
Wei, Zhicheng [4 ]
Yuan, Nicholas Jing [4 ]
机构
[1] Chinese Acad Sci, Inst Software, Chinese Informat Proc Lab, Beijing, Peoples R China
[2] Chinese Acad Sci, Inst Software, State Key Lab Comp Sci, Beijing, Peoples R China
[3] Univ Chinese Acad Sci, Beijing, Peoples R China
[4] Huawei Cloud & AI, Shenzhen, Peoples R China
来源
THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE | 2021年 / 35卷
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Denoising is the essential step for distant supervision based named entity recognition. Previous denoising methods are mostly based on instance-level confidence statistics, which ignore the variety of the underlying noise distribution on different datasets and entity types. This makes them difficult to be adapted to high noise rate settings. In this paper, we propose Hypergeometric Learning (HGL), a denoising algorithm for distantly supervised NER that takes both noise distribution and instance-level confidence into consideration. Specifically, during neural network training, we naturally model the noise samples in each batch following a hypergeometric distribution parameterized by the noise-rate. Then each instance in the batch is regarded as either correct or noisy one according to its label confidence derived from previous training step, as well as the noise distribution in this sampled batch. Experiments show that HGL can effectively denoise the weakly-labeled data retrieved from distant supervision, and therefore results in significant improvements on the trained models.
引用
收藏
页码:14481 / 14488
页数:8
相关论文
共 25 条
[1]  
[Anonymous], 2019, ACL
[2]  
[Anonymous], 2019, ACL
[3]  
Ben Noach M, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P31
[4]  
Chiu J.P.C., 2016, T ASS COMPUTAT LING, V4, P357, DOI [DOI 10.1162/TACLA00104, 10.1162/tacl_a_00104]
[5]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[6]  
Gardner M, 2018, NLP OPEN SOURCE SOFTWARE (NLP-OSS), P1
[7]  
Ju M., 2018, P 2018 C N AM CHAPT, P1446
[8]  
Katiyar A, 2018, P C N AM CHAPT ASS C, DOI [DOI 10.18653/V1/N18-1079, 10.18653/v1/n18-1079]
[9]  
King DB, 2015, ACS SYM SER, V1214, P1, DOI 10.1021/bk-2015-1214.ch001
[10]  
Lample G., 2016, P NAACL, P260