Self-supervised Bidirectional Prompt Tuning for Entity-enhanced Pre-trained Language Model

被引:0
作者
Zou, Jiaxin [1 ]
Xu, Xianghong [1 ]
Hou, Jiawei [2 ]
Yang, Qiang [2 ]
Zheng, Hai-Tao [1 ,3 ]
机构
[1] Tsinghua Univ, Shenzhen Int Grad Sch, Shenzhen 518055, Peoples R China
[2] Weixin Grp, Dept Search & Applicat, Tencent, Peoples R China
[3] Pengcheng Lab, Shenzhen 518055, Peoples R China
来源
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN | 2023年
基金
中国国家自然科学基金;
关键词
D O I
10.1109/IJCNN54540.2023.10192045
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the promotion of the pre-training paradigm, researchers are increasingly focusing on injecting external knowledge, such as entities and triplets from knowledge graphs, into pre-trained language models (PTMs) to improve their understanding and logical reasoning abilities. This results in significant improvements in natural language understanding and generation tasks and some level of interpretability. In this paper, we propose a novel two-stage entity knowledge enhancement pipeline for Chinese pre-trained models based on "bidirectional" prompt tuning. The pipeline consists of a "forward" stage, in which we construct fine-grained entity type prompt templates to boost PTMs injected with entity knowledge, and a "backward" stage, where the trained templates are used to generate type-constrained context-dependent negative samples for contrastive learning. Experiments on six classification tasks in the Chinese Language Understanding Evaluation (CLUE) benchmark demonstrate that our approach significantly improves upon the baseline results in most datasets, particularly those that have a strong reliance on diverse and extensive knowledge.
引用
收藏
页数:8
相关论文
共 50 条
  • [31] Improving Speech Separation with Knowledge Distilled from Self-supervised Pre-trained Models
    Qu, Bowen
    Li, Chenda
    Bai, Jinfeng
    Qian, Yanmin
    2022 13TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2022, : 329 - 333
  • [32] Knowledge Enhanced Pre-trained Language Model for Product Summarization
    Yin, Wenbo
    Ren, Junxiang
    Wu, Yuejiao
    Song, Ruilin
    Liu, Lang
    Cheng, Zhen
    Wang, Sibo
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT II, 2022, 13552 : 263 - 273
  • [33] SSLGuard: A Watermarking Scheme for Self-supervised Learning Pre-trained Encoders<bold> </bold>
    Cong, Tianshuo
    He, Xinlei
    Zhang, Yang
    PROCEEDINGS OF THE 2022 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2022, 2022, : 579 - 593
  • [34] Unstructured Pruning and Low Rank Factorisation of Self-Supervised Pre-Trained Speech Models
    Wang, Haoyu
    Zhang, Wei-Qiang
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2024, 18 (06) : 1046 - 1058
  • [35] Explore the Use of Self-supervised Pre-trained Acoustic Features on Disguised Speech Detection
    Quan, Jie
    Yang, Yingchun
    BIOMETRIC RECOGNITION (CCBR 2021), 2021, 12878 : 483 - 490
  • [36] KNOWLEDGE DISTILLATION FOR NEURAL TRANSDUCERS FROM LARGE SELF-SUPERVISED PRE-TRAINED MODELS
    Yang, Xiaoyu
    Li, Qiujia
    Woodland, Philip C.
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8527 - 8531
  • [37] Self-supervised Learning Based on a Pre-trained Method for the Subtype Classification of Spinal Tumors
    Jiao, Menglei
    Liu, Hong
    Yang, Zekang
    Tian, Shuai
    Ouyang, Hanqiang
    Li, Yuan
    Yuan, Yuan
    Liu, Jianfang
    Wang, Chunjie
    Lang, Ning
    Jiang, Liang
    Yuan, Huishu
    Qian, Yueliang
    Wang, Xiangdong
    COMPUTATIONAL MATHEMATICS MODELING IN CANCER ANALYSIS, CMMCA 2022, 2022, 13574 : 58 - 67
  • [38] ASiT-CRNN: A method for sound event detection with fine-tuning of self-supervised pre-trained ASiT-based model
    Zheng, Yueyang
    Zhang, Ruikun
    Atito, Sara
    Yang, Shuguo
    Wang, Wenwu
    Mei, Yiduo
    DIGITAL SIGNAL PROCESSING, 2025, 160
  • [39] Mitigating Backdoor Attacks in Pre-Trained Encoders via Self-Supervised Knowledge Distillation
    Bie, Rongfang
    Jiang, Jinxiu
    Xie, Hongcheng
    Guo, Yu
    Miao, Yinbin
    Jia, Xiaohua
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2024, 17 (05) : 2613 - 2625
  • [40] Prediction of MASH features from liver biopsy images using a pre-trained self-supervised learning model
    Wang, Yang
    Vyawahare, Saurabh
    McNeil, Carson
    Loo, Jessica
    Robbins, Marc
    Goldenberg, Roman
    JOURNAL OF HEPATOLOGY, 2024, 80 : S592 - S592