Prompt tuning discriminative language models for hierarchical text classification

被引:0
|
作者
du Toit, Jaco [1 ,2 ]
Dunaiski, Marcel [1 ,2 ]
机构
[1] Stellenbosch Univ, Dept Math Sci, Comp Sci Div, Stellenbosch, South Africa
[2] Stellenbosch Univ, Sch Data Sci & Computat Thinking, Stellenbosch, South Africa
来源
NATURAL LANGUAGE PROCESSING | 2024年
关键词
Large language models; discriminative language models; hierarchical text classification; prompt tuning;
D O I
10.1017/nlp.2024.51
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hierarchical text classification (HTC) is a natural language processing task which aims to categorise a text document into a set of classes from a hierarchical class structure. Recent approaches to solve HTC tasks focus on leveraging pre-trained language models (PLMs) and the hierarchical class structure by allowing these components to interact in various ways. Specifically, the Hierarchy-aware Prompt Tuning (HPT) method has proven to be effective in applying the prompt tuning paradigm to Bidirectional Encoder Representations from Transformers (BERT) models for HTC tasks. Prompt tuning aims to reduce the gap between the pre-training and fine-tuning phases by transforming the downstream task into the pre-training task of the PLM. Discriminative PLMs, which use a replaced token detection (RTD) pre-training task, have also shown to perform better on flat text classification tasks when using prompt tuning instead of vanilla fine-tuning. In this paper, we propose the Hierarchy-aware Prompt Tuning for Discriminative PLMs (HPTD) approach which injects the HTC task into the RTD task used to pre-train discriminative PLMs. Furthermore, we make several improvements to the prompt tuning approach of discriminative PLMs that enable HTC tasks to scale to much larger hierarchical class structures. Through comprehensive experiments, we show that our method is robust and outperforms current state-of-the-art approaches on two out of three HTC benchmark datasets.
引用
收藏
页数:18
相关论文
共 50 条
  • [21] Judicial Text Relation Extraction Based on Prompt Tuning
    Chen, Xue
    Li, Yi
    Fan, Shuhuan
    Hou, Mengshu
    2024 2ND ASIA CONFERENCE ON COMPUTER VISION, IMAGE PROCESSING AND PATTERN RECOGNITION, CVIPPR 2024, 2024,
  • [22] Affective Prompt-Tuning-Based Language Model for Semantic-Based Emotional Text Generation
    Gu, Zhaodong
    He, Kejing
    INTERNATIONAL JOURNAL ON SEMANTIC WEB AND INFORMATION SYSTEMS, 2024, 20 (01)
  • [23] Research on fine-tuning strategies for text classification in the aquaculture domain by combining deep learning and large language models
    Zhenglin Li
    Sijia Zhang
    Peirong Cao
    Jiaqi Zhang
    Zongshi An
    Aquaculture International, 2025, 33 (4)
  • [24] An enhanced few-shot text classification approach by integrating topic modeling and prompt-tuning
    Zhang, Yinghui
    Xu, Yichun
    Dong, Fangmin
    NEUROCOMPUTING, 2025, 617
  • [25] Few-Shot Text Classification with an Efficient Prompt Tuning Method in Meta-Learning Framework
    Lv, Xiaobao
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2024, 38 (03)
  • [26] Experiments with hierarchical text classification
    Granitzer, M
    Auer, P
    PROCEEDINGS OF THE NINTH IASTED INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING, 2005, : 177 - 182
  • [27] Prompt Engineering or Fine-Tuning? A Case Study on Phishing Detection with Large Language Models
    Trad, Fouad
    Chehab, Ali
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2024, 6 (01): : 367 - 384
  • [28] G-Prompt: Graphon-based Prompt Tuning for graph classification
    Duan, Yutai
    Liu, Jie
    Chen, Shaowei
    Chen, Liyi
    Wu, Jianhua
    INFORMATION PROCESSING & MANAGEMENT, 2024, 61 (03)
  • [29] Multi-task prompt tuning with soft context sharing for vision-language models
    Ding, Kun
    Wang, Ying
    Liu, Pengzhang
    Yu, Qiang
    Zhang, Haojian
    Xiang, Shiming
    Pan, Chunhong
    NEUROCOMPUTING, 2024, 603
  • [30] Robust scientific text classification using prompt tuning based on data augmentation with L2 regularization
    Shi, Shijun
    Hu, Kai
    Xie, Jie
    Guo, Ya
    Wu, Huayi
    INFORMATION PROCESSING & MANAGEMENT, 2024, 61 (01)