Enhancing rumor detection with data augmentation and generative pre-trained transformer

被引:1
|
作者
Askarizade, Mojgan [1 ]
机构
[1] Ardakan Univ, Fac Engn, Dept Comp Engn, Ardakan, Yazd, Iran
关键词
Fake news detection; Finetuned language model; Neural network classifier; Rumor detection; Generative pre-trained transformer; Data augmentation;
D O I
10.1016/j.eswa.2024.125649
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The advent of social networks has facilitated the rapid dissemination of false information, including rumors, leading to significant societal and individual damages. Extensive research has been dedicated to rumor detection, ranging from machine learning techniques to neural networks. However, the existing methods could not learn the deep concepts of the rumor text to detect the rumor. In addition, imbalanced datasets in the rumor domain reduce the effectiveness of these algorithms. This study addresses this challenge by leveraging the Generative Pre-trained Transformer 2 (GPT-2) model to generate rumor-like texts, thus creating a balanced dataset. Subsequently, a novel approach for classifying rumor texts is proposed by modifying the GPT-2 model. We compare our results with state-of-art machine learning and deep learning methods as well as pretrained models on the PHEME, Twitter15, and Twitter16 datasets. Our findings demonstrate that the proposed model, implementing advanced artificial intelligence techniques, has improved accuracy and F-measure in the application of detecting rumors compared to previous methods.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Using the Chat Generative Pre-trained Transformer in academic writing in health: a scoping review
    Costa, Isabelle Cristinne Pinto
    do Nascimento, Murilo Cesar
    Treviso, Patricia
    Chini, Lucelia Terra
    Roza, Bartira de Aguiar
    Barbosa, Sayonara De Fatima Faria
    Mendes, Karina Dal Sasso
    REVISTA LATINO-AMERICANA DE ENFERMAGEM, 2024, 32
  • [42] Performance of Chat Generative Pre-Trained Transformer on Personal Review of Learning in Obstetrics and Gynecology
    Cohen, Adam
    Burns, Jersey
    Gabra, Martina
    Gordon, Alex
    Deebel, Nicholas
    Terlecki, Ryan
    Woodburn, Katherine L.
    SOUTHERN MEDICAL JOURNAL, 2025, 118 (02) : 102 - 105
  • [43] Text data augmentation and pre-trained Language Model for enhancing text classification of low-resource languages
    Ziyaden, Atabay
    Yelenov, Amir
    Hajiyev, Fuad
    Rustamov, Samir
    Pak, Alexandr
    PEERJ COMPUTER SCIENCE, 2024, 10
  • [44] Chemformer: a pre-trained transformer for computational chemistry
    Irwin, Ross
    Dimitriadis, Spyridon
    He, Jiazhen
    Bjerrum, Esben Jannik
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2022, 3 (01):
  • [45] PART: Pre-trained Authorship Representation Transformer
    Huertas-Tato, Javier
    Martin, Alejandro
    Camacho, David
    HUMAN-CENTRIC COMPUTING AND INFORMATION SCIENCES, 2024, 14
  • [46] Deep Compression of Pre-trained Transformer Models
    Wang, Naigang
    Liu, Chi-Chun
    Venkataramani, Swagath
    Sen, Sanchari
    Chen, Chia-Yu
    El Maghraoui, Kaoutar
    Srinivasan, Vijayalakshmi
    Chang, Leland
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [47] Integrally Pre-Trained Transformer Pyramid Networks
    Tian, Yunjie
    Xie, Lingxi
    Wang, Zhaozhi
    Wei, Longhui
    Zhang, Xiaopeng
    Jiao, Jianbin
    Wang, Yaowei
    Tian, Qi
    Ye, Qixiang
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 18610 - 18620
  • [48] Generative Pre-trained Transformers for Coding Text Data? An Analysis with Classroom Orchestration Data
    Amarasinghe, Ishari
    Marques, Francielle
    Ortiz-Beltran, Ariel
    Hernandez-Leo, Davinia
    RESPONSIVE AND SUSTAINABLE EDUCATIONAL FUTURES, EC-TEL 2023, 2023, 14200 : 32 - 43
  • [49] Pre-trained StyleGAN Based Data Augmentation for Small Sample Brain CT Motion Artifacts Detection
    Su, Kang
    Zhou, Erning
    Sun, Xiaoyu
    Wang, Che
    Yu, Dan
    Luo, Xianlu
    ADVANCED DATA MINING AND APPLICATIONS, 2020, 12447 : 339 - 346
  • [50] GPT-NAS: Neural Architecture Search Meets Generative Pre-Trained Transformer Model
    Yu, Caiyang
    Liu, Xianggen
    Wang, Yifan
    Liu, Yun
    Feng, Wentao
    Deng, Xiong
    Tang, Chenwei
    Lv, Jiancheng
    BIG DATA MINING AND ANALYTICS, 2025, 8 (01): : 45 - 64