Analyzing Robustness of Automatic Scientific Claim Verification Tools against Adversarial Rephrasing Attacks

被引:0
|
作者
Layne, Janet [1 ]
Ratul, Qudrat e. alahy [1 ]
Serra, Edoardo [1 ]
Jajodia, Sushil [2 ]
机构
[1] Boise State Univ, Dept Comp Sci, Boise, ID USA
[2] George Mason Univ, Ctr Secure Informat Syst, Fairfax, VA 22030 USA
基金
美国国家科学基金会;
关键词
Neural networks; adversarial attack; scientific claim verification;
D O I
10.1145/3663481
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The coronavirus pandemic has fostered an explosion of misinformation about the disease, including the risk and effectiveness of vaccination. AI tools for automatic Scientific Claim Verification (SCV) can be crucial to defeat misinformation campaigns spreading through social media channels. However, over the past years, many concerns have been raised about the robustness of AIto adversarial attacks, and the field of automatic SCV is not exempt. The risk is that such SCV tools may reinforce and legitimize the spread of fake scientific claims rather than refute them. This article investigates the problem of generating adversarial attacks for SCV tools and shows that it is far more difficult than the generic NLP adversarial attack problem. The current NLP adversarial attack generators, when applied to SCV, often generate modified claims with entirely different meaning from the original. Even when the meaning is preserved, the modification of the generated claim is too simplistic (only a single word is changed), leaving many weaknesses of the SCV tools undiscovered. We propose T5-ParEvo, an iterative evolutionary attack generator, that is able to generate more complex and creative attacks while better preserving the semantics of the original claim. Using detailed quantitative and qualitative analyses, we demonstrate the efficacy of T5-ParEvo in comparison with existing attack generators.
引用
收藏
页数:32
相关论文
共 26 条
  • [11] Multi-resolution training improves robustness against adversarial attacks
    Shuaiang Rong
    Emadeldeen Hamdan
    Ahmet Enis Cetin
    Signal, Image and Video Processing, 2025, 19 (6)
  • [12] Sparse Adversarial Attacks against DL-Based Automatic Modulation Classification
    Jiang, Zenghui
    Zeng, Weijun
    Zhou, Xingyu
    Feng, Peilun
    Chen, Pu
    Yin, Shenqian
    Han, Changzhi
    Li, Lin
    ELECTRONICS, 2023, 12 (18)
  • [13] Enhancing Robustness of Malware Detection Model Against White Box Adversarial Attacks
    Singhal, Riya
    Soni, Meet
    Bhatt, Shruti
    Khorasiya, Manav
    Jinwala, Devesh C.
    DISTRIBUTED COMPUTING AND INTELLIGENT TECHNOLOGY, ICDCIT 2023, 2023, 13776 : 181 - 196
  • [14] Towards Query-Efficient Adversarial Attacks Against Automatic Speech Recognition Systems
    Wang, Qian
    Zheng, Baolin
    Li, Qi
    Shen, Chao
    Ba, Zhongjie
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 : 896 - 908
  • [15] From environmental sound representation to robustness of 2D CNN models against adversarial attacks
    Esmaeilpour, Mohammad
    Cardinal, Patrick
    Koerich, Alessandro Lameiras
    APPLIED ACOUSTICS, 2022, 195
  • [16] ECG-ATK-GAN: Robustness Against Adversarial Attacks on ECGs Using Conditional Generative Adversarial Networks
    Hossain, Khondker Fariha
    Kamran, Sharif Amit
    Tavakkoli, Alireza
    Ma, Xingjun
    APPLICATIONS OF MEDICAL ARTIFICIAL INTELLIGENCE, AMAI 2022, 2022, 13540 : 68 - 78
  • [17] ROBUSTNESS-AWARE FILTER PRUNING FOR ROBUST NEURAL NETWORKS AGAINST ADVERSARIAL ATTACKS
    Lim, Hyuntak
    Roh, Si-Dong
    Park, Sangki
    Chung, Ki-Seok
    2021 IEEE 31ST INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2021,
  • [18] Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW
    Villegas-Ch, William
    Jaramillo-Alcazar, Angel
    Lujan-Mora, Sergio
    BIG DATA AND COGNITIVE COMPUTING, 2024, 8 (01)
  • [19] Enhancing the robustness of vision transformer defense against adversarial attacks based on squeeze-and-excitation module
    Chang, YouKang
    Zhao, Hong
    Wang, Weijie
    PEERJ COMPUTER SCIENCE, 2023, 9
  • [20] Robustness of Spiking Neural Networks Based on Time-to-First-Spike Encoding Against Adversarial Attacks
    Nomura, Osamu
    Sakemi, Yusuke
    Hosomi, Takeo
    Morie, Takashi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2022, 69 (09) : 3640 - 3644