Mutual Learning Prototype Network for Few-Shot Text Classification

被引:0
|
作者
Liu, Jun [1 ]
Qin, Xiaorui [1 ]
Tao, Jian [1 ]
Dong, Hongfei [1 ]
Li, Xiaoxu [2 ]
机构
[1] China Aero-polytechnology Establishment, Beijing
[2] College of Computer and Communieation, Lanzhou University of Technology, Lanzhou
来源
Beijing Youdian Daxue Xuebao/Journal of Beijing University of Posts and Telecommunications | 2024年 / 47卷 / 03期
关键词
artificial intelligence; few-shot learning; mutual learning; prototype network; text classification;
D O I
10.13190/j.jbupt.2023-114
中图分类号
学科分类号
摘要
Existing methods for few-shot text classification usually rely only on a single prototype for training and inference, which is susceptible to noise and other factors, resulting in insufficient generalization ability. In response to this, a Mutual Learning-Prototype Network for few-shot text classification is proposed. On the basis of retaining the existing algorithm to compute the prototype directly by text embedding features, this paper introduces the BERT network, which inputs the text embedding features into BERT to generate a new prototype. Then, using the mutual learning algorithm, the two prototypes are mutually constrained and knowledge is exchanged to filter out the inaccurate semantic information. This process aims to enhance the feature extraction capability of the model and improve the classification accuracy by joint decision making of the two prototypes. The effectiveness of our proposed approach has been confirmed by the experimental results on a few-shot text classification dataset. The results demonstrate that, on the FewRel dataset, our method improves the accuracy by 2. 97% in the 5-way 1-shot experiment compared to the current optimal method, and by 1. 99% in the 5-way 5-shot experiment. © 2024 Beijing University of Posts and Telecommunications. All rights reserved.
引用
收藏
页码:30 / 35
页数:5
相关论文
共 18 条
  • [1] MINAEE S, KALCHBRENNER N, CAMBRIA E, Et al., Deep learning——based text classification
  • [2] a comprehensive review, ACM Computing Surveys, 54, 3, pp. 1-40, (2021)
  • [3] SNELL J, SWERSKY K, ZEMEL R., Prototypical networks for few-shot learning [J], Advances in Neural Information Processing Systems, 30, pp. 4077-4087, (2017)
  • [4] SUNG F, YANG Y, ZHANG L, Et al., Learning to compare: relation network for few-shot learning, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1199-1208, (2018)
  • [5] ZHOU Z, QIU X, XIE J, Et al., Binocular mutual learning for improving few-shot classification [C], Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 8402-8411, (2021)
  • [6] HUANG Y H, HE Y, YUAN Y J, Et al., Stylizednerf: consistent 3D scene stylization as stylized nerf via 2D-3D mutual learning, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 18342-18352, (2022)
  • [7] DEVLIN J, CHANG M, LEE K, Et al., BERT
  • [8] pretrain-ing of deep bidirectional transformers for language understanding, Proceedings of NAACL-HLT, 1, pp. 4171-4186, (2019)
  • [9] YANG F, WANG R, CHEN X., SEGA
  • [10] semantic guided attention on visual prototype for few-shot learning [C], Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 1056-1066, (2022)