Analyzing Robustness of Automatic Scientific Claim Verification Tools against Adversarial Rephrasing Attacks

被引:0
|
作者
Layne, Janet [1 ]
Ratul, Qudrat e. alahy [1 ]
Serra, Edoardo [1 ]
Jajodia, Sushil [2 ]
机构
[1] Boise State Univ, Dept Comp Sci, Boise, ID USA
[2] George Mason Univ, Ctr Secure Informat Syst, Fairfax, VA 22030 USA
基金
美国国家科学基金会;
关键词
Neural networks; adversarial attack; scientific claim verification;
D O I
10.1145/3663481
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The coronavirus pandemic has fostered an explosion of misinformation about the disease, including the risk and effectiveness of vaccination. AI tools for automatic Scientific Claim Verification (SCV) can be crucial to defeat misinformation campaigns spreading through social media channels. However, over the past years, many concerns have been raised about the robustness of AIto adversarial attacks, and the field of automatic SCV is not exempt. The risk is that such SCV tools may reinforce and legitimize the spread of fake scientific claims rather than refute them. This article investigates the problem of generating adversarial attacks for SCV tools and shows that it is far more difficult than the generic NLP adversarial attack problem. The current NLP adversarial attack generators, when applied to SCV, often generate modified claims with entirely different meaning from the original. Even when the meaning is preserved, the modification of the generated claim is too simplistic (only a single word is changed), leaving many weaknesses of the SCV tools undiscovered. We propose T5-ParEvo, an iterative evolutionary attack generator, that is able to generate more complex and creative attacks while better preserving the semantics of the original claim. Using detailed quantitative and qualitative analyses, we demonstrate the efficacy of T5-ParEvo in comparison with existing attack generators.
引用
收藏
页数:32
相关论文
共 26 条
  • [1] Robustness Against Adversarial Attacks Using Dimensionality
    Chattopadhyay, Nandish
    Chatterjee, Subhrojyoti
    Chattopadhyay, Anupam
    SECURITY, PRIVACY, AND APPLIED CRYPTOGRAPHY ENGINEERING, SPACE 2021, 2022, 13162 : 226 - 241
  • [2] ADVERSARIAL ATTACKS ON SPOOFING COUNTERMEASURES OF AUTOMATIC SPEAKER VERIFICATION
    Liu, Songxiang
    Wu, Haibin
    Lee, Hung-yi
    Meng, Helen
    2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019), 2019, : 312 - 319
  • [3] Analyzing the Robustness of Deep Learning Against Adversarial Examples
    Zhao, Jun
    2018 56TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2018, : 1060 - 1064
  • [4] Investigating Robustness of Adversarial Samples Detection for Automatic Speaker Verification
    Li, Xu
    Li, Na
    Zhong, Jinghua
    Wu, Xixin
    Liu, Xunying
    Su, Dan
    Yu, Dong
    Meng, Helen
    INTERSPEECH 2020, 2020, : 1540 - 1544
  • [5] Enhancing Model Robustness Against Adversarial Attacks with an Anti-adversarial Module
    Qin, Zhiquan
    Liu, Guoxing
    Lin, Xianming
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT IX, 2024, 14433 : 66 - 78
  • [6] Defending Against Adversarial Attacks in Speaker Verification Systems
    Chang, Li-Chi
    Chen, Zesheng
    Chen, Chao
    Wang, Guoping
    Bi, Zhuming
    2021 IEEE INTERNATIONAL PERFORMANCE, COMPUTING, AND COMMUNICATIONS CONFERENCE (IPCCC), 2021,
  • [7] On the Robustness of Intrusion Detection Systems for Vehicles Against Adversarial Attacks
    Choi, Jeongseok
    Kim, Hyoungshick
    INFORMATION SECURITY APPLICATIONS, 2021, 13009 : 39 - 50
  • [8] Relative Robustness of Quantized Neural Networks Against Adversarial Attacks
    Duncan, Kirsty
    Komendantskaya, Ekaterina
    Stewart, Robert
    Lones, Michael
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [9] Towards Unified Robustness Against Both Backdoor and Adversarial Attacks
    Niu, Zhenxing
    Sun, Yuyao
    Miao, Qiguang
    Jin, Rong
    Hua, Gang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 7589 - 7605
  • [10] Chaotic neural network quantization and its robustness against adversarial attacks
    Osama, Alaa
    Gadallah, Samar I.
    Said, Lobna A.
    Radwan, Ahmed G.
    Fouda, Mohammed E.
    KNOWLEDGE-BASED SYSTEMS, 2024, 286