Negative-supervised capsule graph neural network for few-shot text classification

被引:1
作者
Ding, Ling [1 ]
Chen, Xiaojun [1 ]
Xiang, Yang [1 ]
机构
[1] Tongji Univ, Comp Sci & Technol Dept, 4800 Caoan Highway, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Graph neural networks; negative supervision; dynamic routing; few-shot learning;
D O I
10.3233/JIFS-210795
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-shot text classification aims to learn a classifier from very few labeled text data. Existing studies on this topic mainly adopt prototypical networks and focus on interactive information between support set and query instances to learn generalized class prototypes. However, in the process of encoding, these methods only pay attention to the matching information between support set and query instances, and ignore much useful information about intra-class similarity and inter-class dissimilarity between all support samples. Therefore, in this paper we propose a negative-supervised capsule graph neural network (NSCGNN) which explicitly takes use of the similarity and dissimilarity between samples to make the text representations of the same type closer with each other and the ones of different types farther away, leading to representative and discriminative class prototypes. We firstly construct a graph to obtain text representations in the form of node capsules, where both intra-cluster similarity and inter-cluster dissimilarity between all samples are explored with information aggregation and negative supervision. Then, in order to induce generalized class prototypes based on those node capsules obtained from graph neural network, the dynamic routing algorithm is utilized in our model. Experimental results demonstrate the effectiveness of our proposed NSCGNN model, which outperforms existing few-shot approaches on three benchmark datasets.
引用
收藏
页码:6875 / 6887
页数:13
相关论文
共 50 条
  • [1] Bao Y., 2020, P 8 INT C LEARN REPR
  • [2] Bogin B, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P4560
  • [3] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [4] Donahue J, 2014, PR MACH LEARN RES, V32
  • [5] Finn C, 2017, PR MACH LEARN RES, V70
  • [6] Gao TY, 2019, AAAI CONF ARTIF INTE, P6407
  • [7] Garcia V., 2018, 6 INT C LEARN REPR I, P1, DOI 10.48550/arXiv.1711.04043
  • [8] Geng R., 2019, Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), P3895
  • [9] Geng RY, 2020, 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), P1087
  • [10] Gori M, 2005, IEEE IJCNN, P729