Dynamic matching-prototypical learning for noisy few-shot relation classification

被引:0
作者
Bi, Haijia [1 ,3 ]
Peng, Tao [1 ,2 ,3 ]
Han, Jiayu [4 ]
Cui, Hai [1 ,3 ]
Liu, Lu [1 ,2 ,3 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun 130012, Jilin, Peoples R China
[2] Jilin Univ, Coll Software, Changchun 130012, Jilin, Peoples R China
[3] Minist Educ, Key Lab Symbol Computat & Knowledge Engineer, Changchun 130012, Jilin, Peoples R China
[4] Univ Washington, Dept Linguist, Seattle, WA 98195 USA
基金
中国国家自然科学基金;
关键词
Information extraction; Relation classification; Few-shot learning; Dynamic prototype; Dimensional attention; NETWORKS;
D O I
10.1016/j.knosys.2024.112888
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-shot relation classification (FSRC) targets at tackling the long-tail relation problem by identifying relations between entity pairs with few labeled instances. Most of exiting approaches are difficult to deal with text noise in the real world and complex scenes owing to the few labeled sample instances available for learning and the diversity of text expressions. Moreover, they are constrained by limited capacity to extract text features, thus ignoring the proximity of similar relations and the distinctiveness of different relations in the semantic space. Herein, we propose an innovative D ynamic M atching-Prototypical N etwork (DMPN), which equips with three mechanisms for noisy FSRC. In particular, DMPN adopts an I nstance-Guided M atching (IGM) module and a D imension-Guided A ttention (DGA) module to eliminate the deviation of prototype calculation, and dynamically adjusts the contributions of different instances and dimensional features in the support set. Additionally, a two-stage R elation-Aware T raining (RAT) method is designed to focus on noisy relations and improve the discrimination of relation representations. Extensive experiments conduct on the FewRel dataset demonstrate that our model outperform other competitive baselines. Specifically, in the noiseless scenario and under three different noise rate settings, DMPN achieves an average accuracy of 84.71%/80.81%/74.43%/66.81% respectively, which improves from 2.26% to 6.38% on average compared to all baseline methods.
引用
收藏
页数:13
相关论文
共 50 条
[1]  
[Anonymous], 2015, EMNLP
[2]  
[Anonymous], 2014, P COLING 2014 25 INT
[3]  
[Anonymous], 2015, PROC C EMPIRICAL MET
[4]   Predict, pretrained, select and answer: Interpretable and scalable complex question answering over knowledge basest [J].
Cao, Xing ;
Liu, Yun ;
Sun, Feng .
KNOWLEDGE-BASED SYSTEMS, 2023, 278
[5]  
Chen JF, 2022, AAAI CONF ARTIF INTE, P10492
[6]  
Chen XD, 2023, PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, P7409
[7]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[8]   Attribute Group Editing for Reliable Few-shot Image Generation [J].
Ding, Guanqi ;
Han, Xinzhe ;
Wang, Shuhui ;
Wu, Shuzhe ;
Jin, Xin ;
Tu, Dandan ;
Huang, Qingming .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :11184-11193
[9]  
Ding N, 2023, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, P895
[10]  
Dong Q., 2023, CoRR 2301.00234