Pre-trained language models for keyphrase prediction: A review

被引:2
|
作者
Umair, Muhammad [1 ]
Sultana, Tangina [1 ,2 ]
Lee, Young-Koo [1 ]
机构
[1] Kyung Hee Univ, Dept Comp Sci & Engn, Global Campus, Yongin, South Korea
[2] Hajee Mohammad Danesh Sci & Technol Univ, Dept Elect & Commun Engn, Dinajpur, Bangladesh
来源
ICT EXPRESS | 2024年 / 10卷 / 04期
关键词
Keyphrases; Keyphrase extraction; Keyphrase generation; Pre-trained language models; Natural language processing; Large language models; Review; EXTRACTION;
D O I
10.1016/j.icte.2024.05.015
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Keyphrase Prediction (KP) is essential for identifying keyphrases in a document that can summarize its content. However, recent Natural Language Processing (NLP) advances have developed more efficient KP models using deep learning techniques. The limitation of a comprehensive exploration jointly both keyphrase extraction and generation using pre-trained language models spotlights a critical gap in the literature, compelling our survey paper to bridge this deficiency and offer a unified and in-depth analysis to address limitations in previous surveys. This paper extensively examines the topic of pre-trained language models for keyphrase prediction (PLM-KP), which are trained on large text corpora via different learning (supervisor, unsupervised, semi-supervised, and self-supervised) techniques, to provide respective insights into these two types of tasks in NLP, precisely, Keyphrase Extraction (KPE) and Keyphrase Generation (KPG). We introduce appropriate taxonomies for PLM-KPE and KPG to highlight these two main tasks of NLP. Moreover, we point out some promising future directions for predicting keyphrases. (c) 2024 The Author(s). Published by Elsevier B.V. on behalf of The Korean Institute of Communications and Information Sciences. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页码:871 / 890
页数:20
相关论文
共 50 条
  • [21] Topic-Attentive Encoder-Decoder with Pre-Trained Language Model for Keyphrase Generation
    Zhou, Cangqi
    Shang, Jinling
    Zhang, Jing
    Li, Qianmu
    Hu, Dimming
    2021 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2021), 2021, : 1529 - 1534
  • [22] Rethinking Model Selection and Decoding for Keyphrase Generation with Pre-trained Sequence-to-Sequence Models
    Wu, Di
    Ahmad, Wasi Uddin
    Chang, Kai-Wei
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 6642 - 6658
  • [23] Vulnerability Analysis of Continuous Prompts for Pre-trained Language Models
    Li, Zhicheng
    Shi, Yundi
    Sheng, Xuan
    Yin, Changchun
    Zhou, Lu
    Li, Piji
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT IX, 2023, 14262 : 508 - 519
  • [24] Adapting Pre-trained Language Models to Rumor Detection on Twitter
    Slimi, Hamda
    Bounhas, Ibrahim
    Slimani, Yahya
    JOURNAL OF UNIVERSAL COMPUTER SCIENCE, 2021, 27 (10) : 1128 - 1148
  • [25] ProSide: Knowledge Projector and Sideway for Pre-trained Language Models
    He, Chaofan
    Lu, Gewei
    Shen, Liping
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT II, NLPCC 2024, 2025, 15360 : 56 - 68
  • [26] Exploring Pre-trained Language Models for Vocabulary Alignment in the UMLS
    Hao, Xubing
    Abeysinghe, Rashmie
    Shi, Jay
    Cui, Licong
    ARTIFICIAL INTELLIGENCE IN MEDICINE, PT I, AIME 2024, 2024, 14844 : 273 - 278
  • [27] Focused Contrastive Loss for Classification With Pre-Trained Language Models
    He, Jiayuan
    Li, Yuan
    Zhai, Zenan
    Fang, Biaoyan
    Thorne, Camilo
    Druckenbrodt, Christian
    Akhondi, Saber
    Verspoor, Karin
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (07) : 3047 - 3061
  • [28] Quantifying Gender Bias in Arabic Pre-Trained Language Models
    Alrajhi, Wafa
    Al-Khalifa, Hend S.
    Al-Salman, Abdulmalik S.
    IEEE ACCESS, 2024, 12 : 77406 - 77420
  • [29] Discrimination Bias Detection Through Categorical Association in Pre-Trained Language Models
    Dusi, Michele
    Arici, Nicola
    Gerevini, Alfonso Emilio
    Putelli, Luca
    Serina, Ivan
    IEEE ACCESS, 2024, 12 : 162651 - 162667
  • [30] Porter 6: Protein Secondary Structure Prediction by Leveraging Pre-Trained Language Models (PLMs)
    Alanazi, Wafa
    Meng, Di
    Pollastri, Gianluca
    INTERNATIONAL JOURNAL OF MOLECULAR SCIENCES, 2025, 26 (01)