Cost-effective CNNs-based prototypical networks for few-shot relation classification across domains

被引:8
作者
Yin, Gongzhu [1 ]
Wang, Xing [1 ]
Zhang, Hongli [1 ]
Wang, Jinlin [1 ]
机构
[1] Harbin Inst Technol, Sch Cyberspace Sci, Harbin, Heilongjiang, Peoples R China
关键词
Relation classification; Few-shot learning; Domain adaptation; Prototypical network;
D O I
10.1016/j.knosys.2022.109470
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper studies few-shot relation classification under domain shift, which is quite a challenging inductive task in practice. Previous work focusing on few-shot relation classification usually adopted prototypical networks, whose performance dramatically dropped when adapting to diverse domains. Some researches introduced large pretrained language models, which consume massive time and computation resources. To address the above issues, we propose cost-effective CNNs-based prototypical networks in this paper. Specifically, a multichannel encoder (MCE) is adopted to capture general domain invariant features respectively from the entity and the context, then they are aggregated according to relation classes. When encoding the context, we propose an attention mechanism based on the dependency trees of sentences to effectively select helpful grams. To get further improvements, we leverage the unlabeled data from the target domain by pseudo-labeling and introduce a method to select instances with high confidence via information entropy. We conducted experiments on two public datasets: FewRel 2.0 and FewTAC. The results demonstrate that our approaches not only largely enhance the effectiveness of original prototypical networks, but also achieve competitive results with large pretrained models with faster speeds and much fewer computational costs. (C) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:11
相关论文
共 41 条
[1]  
Bunescu R., 2005, P HUM LANG TECHN C C, P724
[2]  
Bunescu R. C., 2005, Proceedings of the 18th International Conference on Neural Information Processing Systems, P171, DOI DOI 10.5555/2976248.2976270
[3]  
Cong X., 2020, LECT NOTES COMPUTER, P624, DOI 10.1007/2F978-3-030-67661-2_37
[4]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[5]  
Fangchao Liu, 2021, Chinese Computational Linguistics: 20th China National Conference, CCL 2021, Proceedings. Lecture Notes in Computer Science, Lecture Notes in Artificial Intelligence (12869), P193, DOI 10.1007/978-3-030-84186-7_13
[6]  
Finn C, 2017, PR MACH LEARN RES, V70
[7]  
Gao TY, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P6250
[8]  
Gao TY, 2019, AAAI CONF ARTIF INTE, P6407
[9]   MICK: A Meta-Learning Framework for Few-shot Relation Classification with Small Training Data [J].
Geng, Xiaoqing ;
Chen, Xiwen ;
Kenny Q Zhu ;
Shen, Libin ;
Zhao, Yinggong .
CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, :415-424
[10]  
Guo ZJ, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P241