Predicting Terms in IS-A Relations with Pre-trained Transformers

被引:0
|
作者
Nikishina, Irina [1 ]
Chernomorchenko, Polina [2 ]
Demidova, Anastasiia [3 ]
Panchenko, Alexander [3 ,4 ]
Biemann, Chris [1 ]
机构
[1] Univ Hamburg, Hamburg, Germany
[2] HSE Univ, Moscow, Russia
[3] Skolkovo Inst Sci & Technol, Moscow, Russia
[4] AIRI, Moscow, Russia
来源
13TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING AND THE 3RD CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, IJCNLP-AACL 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we explore the ability of the generative transformers to predict objects in IS-A (hypo-hypernym) relations. We solve the task for both directions of the relations: we learn to predict hypernyms given the input word and hyponyms, given the input concept and its neighbourhood from the taxonomy. To the best of our knowledge, this is the first paper which provides a comprehensive analysis of transformerbased models for the task of hypernymy extraction. Apart from the standard finetuning of various generative models, we experiment with different input formats and prefixes, zeroand few-shot learning strategies, and generation parameters. Results show that higher performance on both subtasks can be achieved by generative transformers with no additional data (like definitions or lemma names). Such models have phenomenally high abilities at the task given a little training and proper prompts in comparison to specialized rule-based and statistical methods as well as encoder-based transformer models.
引用
收藏
页码:134 / 148
页数:15
相关论文
共 50 条
  • [41] EAPT: An encrypted traffic classification model via adversarial pre-trained transformers
    Zhan, Mingming
    Yang, Jin
    Jia, Dongqing
    Fu, Geyuan
    COMPUTER NETWORKS, 2025, 257
  • [42] Detecting Propaganda Techniques in English News Articles using Pre-trained Transformers
    Abdullah, Malak
    Altiti, Ola
    Obiedat, Rasha
    2022 13TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION SYSTEMS (ICICS), 2022, : 301 - 308
  • [43] CopiFilter: An Auxiliary Module Adapts Pre-trained Transformers for Medical Dialogue Summarization
    Duan, Jiaxin
    Liu, Junfei
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT IV, 2023, 14257 : 99 - 114
  • [44] Math-LLMs: AI Cyberinfrastructure with Pre-trained Transformers for Math Education
    Zhang, Fan
    Li, Chenglu
    Henkel, Owen
    Xing, Wanli
    Baral, Sami
    Heffernan, Neil
    Li, Hai
    INTERNATIONAL JOURNAL OF ARTIFICIAL INTELLIGENCE IN EDUCATION, 2024,
  • [45] Towards a Comprehensive Understanding and Accurate Evaluation of Societal Biases in Pre-Trained Transformers
    Silva, Andrew
    Tambwekar, Pradyumna
    Gombolay, Matthew
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 2383 - 2389
  • [46] Harnessing Generative Pre-Trained Transformers for Construction Accident Prediction with Saliency Visualization
    Yoo, Byunghee
    Kim, Jinwoo
    Park, Seongeun
    Ahn, Changbum R.
    Oh, Taekeun
    APPLIED SCIENCES-BASEL, 2024, 14 (02):
  • [47] BERT-QPP: Contextualized Pre-trained Transformers for Query Performance Prediction
    Arabzadeh, Negar
    Khodabakhsh, Maryam
    Bagheri, Ebrahim
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 2857 - 2861
  • [48] PAIR: Planning and Iterative Refinement in Pre-trained Transformers for Long Text Generation
    Hua, Xinyu
    Wang, Lu
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 781 - 793
  • [49] Learning to Switch off, Switch on, and Integrate Modalities in Large Pre-trained Transformers
    Duseja, Tejas
    Annervaz, K. M.
    Duggani, Jeevithiesh
    Zacharia, Shyam
    Free, Michael
    Dukkipati, Ambedkar
    2024 IEEE 7TH INTERNATIONAL CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL, MIPR 2024, 2024, : 403 - 409
  • [50] ProdRev: A DNN framework for empowering customers using generative pre-trained transformers
    Gupta, Aakash
    Das, Nataraj
    2022 INTERNATIONAL CONFERENCE ON DECISION AID SCIENCES AND APPLICATIONS (DASA), 2022, : 895 - 899