Dual Contrastive Learning for Cross-Domain Named Entity Recognition

被引:1
作者
Xu, Jingyun [1 ]
Yu, Junnan [1 ]
Cai, Yi [1 ]
Chua, Tat-Seng [2 ]
机构
[1] South China Univ Technol, Guangzhou, Peoples R China
[2] Natl Univ Singapore, Singapore City, Singapore
基金
中国国家自然科学基金;
关键词
Named Entity Recognition; Cross-domain; Contrastive Learning; EXTRACTION;
D O I
10.1145/3678879
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Benefiting many information retrieval applications, named entity recognition (NER) has shown impressive progress. Recently, there has been a growing trend to decompose complex NER tasks into two subtasks (e.g., entity span detection (ESD) and entity type classification (ETC), to achieve better performance. Despite the remarkable success, from the perspective of representation, existing methods do not explicitly distinguish non-entities and entities, which may lead to ESD errors. Meanwhile, they do not explicitly distinguish entities with different entity types, which may lead to entity type misclassification. As such, the limited representation abilities may challenge some competitive NER methods, leading to unsatisfactory performance, especially in the low-resource setting (e.g., cross-domain NER). In light of these challenges, we propose to utilize contrastive learning to refine the original chaotic representations and learn the generalized representations for cross- domain NER. In particular, this article proposes a dual contrastive learning model (Dual-CL), which respectively utilizes a token-level contrastive learning module and a sentence-level contrastive learning module to enhance ESD, ETC for cross-domain NER. Empirical results on 10 domain pairs under two different settings show that Dual-CL achieves better performances than compared baselines in terms of several standard metrics. Moreover, we conduct detailed analyses to are presented to better understand each component's effectiveness.
引用
收藏
页数:33
相关论文
共 129 条
[1]  
Caron M, 2020, ADV NEUR IN, V33
[2]   Emerging Properties in Self-Supervised Vision Transformers [J].
Caron, Mathilde ;
Touvron, Hugo ;
Misra, Ishan ;
Jegou, Herve ;
Mairal, Julien ;
Bojanowski, Piotr ;
Joulin, Armand .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :9630-9640
[3]  
Changki Lee, 2007, 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, P799
[4]  
Chen T, 2020, PR MACH LEARN RES, V119
[5]   An Empirical Cross Domain-Specific Entity Recognition with Domain Vector [J].
Chen, Wei ;
Han, Songqiao ;
Huang, Hailiang .
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, :3868-3872
[6]   Exploring Simple Siamese Representation Learning [J].
Chen, Xinlei ;
He, Kaiming .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :15745-15753
[7]   Intent Contrastive Learning for Sequential Recommendation [J].
Chen, Yongjun ;
Liu, Zhiwei ;
Li, Jia ;
McAuley, Julian ;
Xiong, Caiming .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :2172-2182
[8]  
Dai X., 2020, ACL, P5860, DOI [10.18653/V1/2020.ACL-MAIN.520, DOI 10.18653/V1/2020.ACL-MAIN.520]
[9]  
Das SSS, 2022, PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), P6338
[10]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171