PTE: Prompt tuning with ensemble verbalizers

被引:0
|
作者
Liang, Liheng [1 ]
Wang, Guancheng [2 ]
Lin, Cong [2 ]
Feng, Zhuowen [3 ]
机构
[1] Guangdong Ocean Univ, Fac Math & Comp Sci, Zhanjiang 524088, Peoples R China
[2] Guangdong Ocean Univ, Coll Elect & Informat Engn, Zhanjiang 524088, Peoples R China
[3] Guangdong Ocean Univ, Coll Literature & News Commun, Zhanjiang 524088, Peoples R China
关键词
Prompt tuning; Few-shot learning; Text classification; Pre-trained language models;
D O I
10.1016/j.eswa.2024.125600
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prompt tuning has achieved remarkable success in facilitating the performance of Pre-trained Language Models (PLMs) across various downstream NLP tasks, particularly in scenarios with limited downstream data. Reframing tasks as fill-in-the-blank questions represents an effective approach within prompt tuning. However, this approach necessitates the mapping of labels through a verbalizer consisting of one or more label tokens, constrained by manually crafted prompts. Furthermore, most existing automatic crafting methods either introduce external resources or rely solely on discrete or continuous optimization strategies. To address this issue, we have proposed a methodology for optimizing discrete verbalizers based on gradient descent, which we refer to this approach as PTE. This method integrates discrete tokens into verbalizers that can be continuously optimized, combining the distinct advantages of both discrete and continuous optimization strategies. In contrast to prior approaches, ours eschews reliance on prompts generated by other models or prior knowledge, merely augmenting a matrix. This approach boasts remarkable simplicity and flexibility, enabling prompt optimization while preserving the interpretability of output label tokens without constraints imposed by discrete vocabularies. Finally, employing this method in text classification tasks, we observe that PTE achieves results comparable to, if not surpassing, previous methods even under extreme conciseness. This furnishes a simple, intuitive, and efficient solution for automatically constructing verbalizers. Moreover, through quantitative analysis of optimized verbalizers, we uncover that language models likely rely not only on semantic information but also on other features for text classification. This revelation unveils new avenues for future research and model enhancements.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Short text classification with Soft Knowledgeable Prompt-tuning
    Zhu, Yi
    Wang, Ye
    Mu, Jianyuan
    Li, Yun
    Qiang, Jipeng
    Yuan, Yunhao
    Wu, Xindong
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 246
  • [42] Filling in the Blank: Rationale-Augmented Prompt Tuning for TextVQA
    Zeng, Gangyan
    Zhang, Yuan
    Zhou, Yu
    Fang, Bo
    Zhao, Guoqing
    Wei, Xin
    Wang, Weiping
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 1261 - 1272
  • [43] Ontology-based prompt tuning for news article summarization
    Silva, A. R. S.
    Priyadarshana, Y. H. P. P.
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2025, 8
  • [44] COPHTC: CONTRASTIVE LEARNING WITH PROMPT TUNING FOR HIERARCHICAL TEXT CLASSIFICATION
    Cai, Fuhan
    Zhang, Zhongqiang
    Liu, Duo
    Fang, Xiangzhong
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 5400 - 5404
  • [45] Prompt tuning for parameter-efficient medical image segmentation
    Fischer, Marc
    Bartler, Alexander
    Yang, Bin
    MEDICAL IMAGE ANALYSIS, 2024, 91
  • [46] VPN: Variation on Prompt Tuning for Named-Entity Recognition
    Hu, Niu
    Zhou, Xuan
    Xu, Bing
    Liu, Hanqing
    Xie, Xiangjin
    Zheng, Hai-Tao
    APPLIED SCIENCES-BASEL, 2023, 13 (14):
  • [47] Data Augmentation by Prompt Tuning on Natural Language Understanding Tasks
    Wang, Yu-Hao
    Chang, Chia-Ming
    Tsai, Yi-Hang
    Hwang, San-Yih
    2024 11TH INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS-TAIWAN, ICCE-TAIWAN 2024, 2024, : 807 - 808
  • [48] POET: Prompt Offset Tuning for Continual Human Action Adaptation
    Garg, Prachi
    Joseph, K. J.
    Balasubramanian, Vineeth N.
    Camgoz, Necati Cihan
    Wan, Chengde
    King, Kenrick
    Si, Weiguang
    Ma, Shugao
    De La Torre, Fernando
    COMPUTER VISION - ECCV 2024, PT LXIV, 2025, 15122 : 436 - 455
  • [49] Model tuning or prompt Tuning? a study of large language models for clinical concept and relation extraction
    Peng, Cheng
    Yang, Xi
    Smith, Kaleb E.
    Yu, Zehao
    Chen, Aokun
    Bian, Jiang
    Wu, Yonghui
    JOURNAL OF BIOMEDICAL INFORMATICS, 2024, 153
  • [50] Context-aware generative prompt tuning for relation extraction
    Liu, Xiaoyong
    Wen, Handong
    Xu, Chunlin
    Du, Zhiguo
    Li, Huihui
    Hu, Miao
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (12) : 5495 - 5508