Enhancing rumor detection with data augmentation and generative pre-trained transformer

被引:1
|
作者
Askarizade, Mojgan [1 ]
机构
[1] Ardakan Univ, Fac Engn, Dept Comp Engn, Ardakan, Yazd, Iran
关键词
Fake news detection; Finetuned language model; Neural network classifier; Rumor detection; Generative pre-trained transformer; Data augmentation;
D O I
10.1016/j.eswa.2024.125649
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The advent of social networks has facilitated the rapid dissemination of false information, including rumors, leading to significant societal and individual damages. Extensive research has been dedicated to rumor detection, ranging from machine learning techniques to neural networks. However, the existing methods could not learn the deep concepts of the rumor text to detect the rumor. In addition, imbalanced datasets in the rumor domain reduce the effectiveness of these algorithms. This study addresses this challenge by leveraging the Generative Pre-trained Transformer 2 (GPT-2) model to generate rumor-like texts, thus creating a balanced dataset. Subsequently, a novel approach for classifying rumor texts is proposed by modifying the GPT-2 model. We compare our results with state-of-art machine learning and deep learning methods as well as pretrained models on the PHEME, Twitter15, and Twitter16 datasets. Our findings demonstrate that the proposed model, implementing advanced artificial intelligence techniques, has improved accuracy and F-measure in the application of detecting rumors compared to previous methods.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Generative Pre-Trained Transformer (GPT) in Research: A Systematic Review on Data Augmentation
    Sufi, Fahim
    INFORMATION, 2024, 15 (02)
  • [2] Generative Pre-Trained Transformer for Cardiac Abnormality Detection
    Gaudilliere, Pierre Louis
    Sigurthorsdottir, Halla
    Aguet, Clementine
    Van Zaen, Jerome
    Lemay, Mathieu
    Delgado-Gonzalo, Ricard
    2021 COMPUTING IN CARDIOLOGY (CINC), 2021,
  • [3] Adapting Pre-trained Generative Model to Medical Image for Data Augmentation
    Yuan, Zhouhang
    Fang, Zhengqing
    Huang, Zhengxing
    Wu, Fei
    Yao, Yu-Feng
    Li, Yingming
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT V, 2024, 15005 : 79 - 89
  • [4] Enhancing clinical reasoning with Chat Generative Pre-trained Transformer: a practical guide
    Hirosawa, Takanobu
    Shimizu, Taro
    DIAGNOSIS, 2024, 11 (01) : 102 - 105
  • [5] Leveraging Generative Pre-Trained Transformer Models for Standardizing Nursing Data
    Baranwal, Aseem
    Semenov, Alexander
    Salgado, Patricia de Oliveira
    Priola, Karen B.
    Yao, Yingwei
    Keenan, Gail M.
    Macieira, Tamara G. R.
    2024 IEEE 12TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS, ICHI 2024, 2024, : 386 - 391
  • [6] OMPGPT: A Generative Pre-trained Transformer Model for OpenMP
    Chen, Le
    Bhattacharjee, Arijit
    Ahmed, Nesreen
    Hasabnis, Niranjan
    Oren, Gal
    Vo, Vy
    Jannesari, Ali
    EURO-PAR 2024: PARALLEL PROCESSING, PT I, EURO-PAR 2024, 2024, 14801 : 121 - 134
  • [7] Enhancing catalysis studies with chat generative pre-trained transformer (ChatGPT): Conversation with ChatGPT
    Ansari, Navid
    Babaei, Vahid
    Najafpour, Mohammad Mahdi
    DALTON TRANSACTIONS, 2024, 53 (08) : 3534 - 3547
  • [8] The application of Chat Generative Pre-trained Transformer in nursing education
    Liu, Jialin
    Liu, Fan
    Fang, Jinbo
    Liu, Siru
    NURSING OUTLOOK, 2023, 71 (06)
  • [9] Data Augmentation Based on Pre-trained Language Model for Event Detection
    Zhang, Meng
    Xie, Zhiwen
    Liu, Jin
    CCKS 2021 - EVALUATION TRACK, 2022, 1553 : 59 - 68
  • [10] Adapting Pre-trained Language Models to Rumor Detection on Twitter
    Slimi, Hamda
    Bounhas, Ibrahim
    Slimani, Yahya
    JOURNAL OF UNIVERSAL COMPUTER SCIENCE, 2021, 27 (10) : 1128 - 1148