Prompt Tuning on Graph-Augmented Low-Resource Text Classification

被引:0
|
作者
Wen, Zhihao [1 ]
Fang, Yuan [1 ]
机构
[1] Singapore Management Univ, Sch Comp & Informat Syst, Singapore 188065, Singapore
关键词
Tuning; Text categorization; Task analysis; Accuracy; Paints; Oils; Ink; Text classification; graph; low-resource learning; pre-training; prompt;
D O I
10.1109/TKDE.2024.3440068
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text classification is a fundamental problem in information retrieval with many real-world applications, such as predicting the topics of online articles and the categories of e-commerce product descriptions. However, low-resource text classification, with no or few labeled samples, presents a serious concern for supervised learning. Meanwhile, many text data are inherently grounded on a network structure, such as a hyperlink/citation network for online articles, and a user-item purchase network for e-commerce products. These graph structures capture rich semantic relationships, which can potentially augment low-resource text classification. In this paper, we propose a novel model called Graph-Grounded Pre-training and Prompting (G2P2) to address low-resource text classification in a two-pronged approach. During pre-training, we propose three graph interaction-based contrastive strategies to jointly pre-train a graph-text model; during downstream classification, we explore handcrafted discrete prompts and continuous prompt tuning for the jointly pre-trained model to achieve zero- and few-shot classification, respectively. Moreover, we explore the possibility of employing continuous prompt tuning for zero-shot inference. Specifically, we aim to generalize continuous prompts to unseen classes while leveraging a set of base classes. To this end, we extend G2P2 into G2P2 (& lowast;) , hinging on a new architecture of conditional prompt tuning. Extensive experiments on four real-world datasets demonstrate the strength of G2P2 in zero- and few-shot low-resource text classification tasks, and illustrate the advantage of G2P2 (& lowast; )in dealing with unseen classes.
引用
收藏
页码:9080 / 9095
页数:16
相关论文
共 22 条
  • [1] Augmenting Low-Resource Text Classification with Graph-Grounded Pre-training and Prompting
    Wen, Zhihao
    Fang, Yuan
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 506 - 516
  • [2] Domain-Aligned Data Augmentation for Low-Resource and Imbalanced Text Classification
    Stylianou, Nikolaos
    Chatzakou, Despoina
    Tsikrika, Theodora
    Vrochidis, Stefanos
    Kompatsiaris, Ioannis
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2023, PT II, 2023, 13981 : 172 - 187
  • [3] Text data augmentation and pre-trained Language Model for enhancing text classification of low-resource languages
    Ziyaden, Atabay
    Yelenov, Amir
    Hajiyev, Fuad
    Rustamov, Samir
    Pak, Alexandr
    PEERJ COMPUTER SCIENCE, 2024, 10
  • [4] Meta-Prompt: Boosting Whisper's Performance in Low-Resource Speech Recognition
    Chen, Yaqi
    Niu, Tong
    Zhang, Hao
    Zhang, Wenlin
    Qu, Dan
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 3039 - 3043
  • [5] Text Classification Based on Convolutional Neural Networks and Word Embedding for Low-Resource Languages: Tigrinya
    Fesseha, Awet
    Xiong, Shengwu
    Emiru, Eshete Derb
    Diallo, Moussa
    Dahou, Abdelghani
    INFORMATION, 2021, 12 (02) : 1 - 17
  • [6] A Scheme for News Article Classification in a Low-Resource Language
    Yohannes, Hailemariam Mehari
    Amagasa, Toshiyuki
    INFORMATION INTEGRATION AND WEB INTELLIGENCE, IIWAS 2022, 2022, 13635 : 519 - 530
  • [7] Knowledge Transfer for Utterance Classification in Low-Resource Languages
    Smirnov, Andrei
    Mendelev, Valentin
    SPEECH AND COMPUTER, 2016, 9811 : 435 - 442
  • [8] Lightweight Automatic Modulation Classification Based on Efficient Convolution and Graph Sparse Attention in Low-Resource Scenarios
    Cai, Zhuoran
    Wang, Chuan
    Ma, Wenxuan
    Li, Xiangzhen
    Zhou, Ruoyu
    IEEE INTERNET OF THINGS JOURNAL, 2025, 12 (04): : 3629 - 3638
  • [9] Soft Prompt-tuning with Self-Resource Verbalizer for short text streams
    Zhu, Yi
    Wang, Ye
    Li, Yun
    Qiang, Jipeng
    Yuan, Yunhao
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 139
  • [10] ANALYZING ASR PRETRAINING FOR LOW-RESOURCE SPEECH-TO-TEXT TRANSLATION
    Stoian, Mihaela C.
    Bansal, Sameer
    Goldwater, Sharon
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7909 - 7913