CLUR: Uncertainty Estimation for Few-Shot Text Classification with Contrastive Learning

被引:1
|
作者
He, Jianfeng [1 ]
Zhang, Xuchao [2 ]
Lei, Shuo [1 ]
Alhamadani, Abdulaziz [1 ]
Chen, Fanglan [1 ]
Xiao, Bei [3 ]
Lu, Chang-Tien [1 ]
机构
[1] Virginia Tech, Falls Church, VA 22043 USA
[2] Microsoft, Redmond, WA USA
[3] Amer Univ, Washington, DC USA
来源
PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023 | 2023年
基金
美国国家科学基金会;
关键词
Uncertainty estimation; few-shot; pseudo labels; contrastive learning;
D O I
10.1145/3580305.3599276
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Few-shot text classification has extensive application where the sample collection is expensive or complicated. When the penalty for classification errors is high, such as early threat event detection with scarce data, we expect to know "whether we should trust the classification results or reexamine them." This paper investigates the Uncertainty Estimation for Few-shot Text Classification (UEFTC), an unexplored research area. Given limited samples, a UEFTC model predicts an uncertainty score for a classification result, which is the likelihood that the classification result is false. However, many traditional uncertainty estimation models in text classification are unsuitable for implementing a UEFTC model. These models require numerous training samples, whereas the few-shot setting in UEFTC only provides a few or just one support sample for each class in an episode. We propose Contrastive Learning from Uncertainty Relations (CLUR) to address UEFTC. CLUR can be trained with only one support sample for each class with the help of pseudo uncertainty scores. Unlike previous works that manually set the pseudo uncertainty scores, CLUR self-adaptively learns them using our proposed uncertainty relations. Specifically, we explore four model structures in CLUR to investigate the performance of three common-used contrastive learning components in UEFTC and find that two of the components are effective. Experiment results prove that CLUR outperforms six baselines on four datasets, including an improvement of 4.52% AUPR on an RCV1 dataset in a 5-way 1-shot setting. Our code and data split for UEFTC are in https: //github.com/he159ok/CLUR_UncertaintyEst_FewShot_TextCls.
引用
收藏
页码:698 / 710
页数:13
相关论文
共 50 条
  • [31] Few-shot contrastive learning for image classification and its application to insulator identification
    Li, Liang
    Jin, Weidong
    Huang, Yingkun
    APPLIED INTELLIGENCE, 2022, 52 (06) : 6148 - 6163
  • [32] SCL: Self-supervised contrastive learning for few-shot image classification
    Lim, Jit Yan
    Lim, Kian Ming
    Lee, Chin Poo
    Tan, Yong Xuan
    NEURAL NETWORKS, 2023, 165 : 19 - 30
  • [33] Cross-Domain Few-Shot Contrastive Learning for Hyperspectral Images Classification
    Zhang, Suhua
    Chen, Zhikui
    Wang, Dan
    Wang, Z. Jane
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [34] Momentum memory contrastive learning for transfer-based few-shot classification
    Runliang Tian
    Hongmei Shi
    Applied Intelligence, 2023, 53 : 864 - 878
  • [35] Momentum memory contrastive learning for transfer-based few-shot classification
    Tian, Runliang
    Shi, Hongmei
    APPLIED INTELLIGENCE, 2023, 53 (01) : 864 - 878
  • [36] Contrastive Knowledge-Augmented Meta-Learning for Few-Shot Classification
    Subramanyam, Rakshith
    Heimann, Mark
    Jayram, T. S.
    Anirudh, Rushil
    Thiagarajan, Jayaraman J.
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 2478 - 2486
  • [37] Domain-Invariant Few-Shot Contrastive Learning for Hyperspectral Image Classification
    Chen, Wenchen
    Zhang, Yanmei
    Chu, Jianping
    Wang, Xingbo
    Applied Sciences (Switzerland), 2024, 14 (23):
  • [38] Causal representation for few-shot text classification
    Yang, Maoqin
    Zhang, Xuejie
    Wang, Jin
    Zhou, Xiaobing
    APPLIED INTELLIGENCE, 2023, 53 (18) : 21422 - 21432
  • [39] Adversarial training for few-shot text classification
    Croce, Danilo
    Castellucci, Giuseppe
    Basili, Roberto
    INTELLIGENZA ARTIFICIALE, 2020, 14 (02) : 201 - 214
  • [40] Causal representation for few-shot text classification
    Maoqin Yang
    Xuejie Zhang
    Jin Wang
    Xiaobing Zhou
    Applied Intelligence, 2023, 53 : 21422 - 21432