Defense against adversarial attacks via textual embeddings based on semantic associative field

被引:0
|
作者
Huang, Jiacheng [1 ]
Chen, Long [1 ,2 ]
机构
[1] Chongqing Univ Posts & Telecommun, Sch Comp Sci & Technol, Chongqing 400065, Peoples R China
[2] Chongqing Univ Posts & Telecommun, Sch Cyber Secur & Informat Law, Chongqing 400065, Peoples R China
来源
NEURAL COMPUTING & APPLICATIONS | 2024年 / 36卷 / 01期
关键词
Adversarial examples; Natural language processing; Semantic associative field; Word-level;
D O I
10.1007/s00521-023-08946-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks are known to be vulnerable to various types of adversarial attacks, especially word-level attacks, in the field of natural language processing. In recent years, various defense methods are proposed against word-level attacks; however, most of those defense methods only focus on synonyms substitution-based attacks, while word-level attacks are not based on synonym substitution. In this paper, we propose a textual adversarial defense method against word-level adversarial attacks via textual embedding based on the semantic associative field. More specifically, we analyze the reasons why humans can read and understand textual adversarial examples and observe two crucial points: (1) There must be a relation between the original word and the perturbed word or token. (2) Such a kind of relation enables humans to infer original words, while humans have the ability to associations. Motivated by this, we introduce the concept of semantic associative field and propose a new defense method by building a robust word embedding, that is, we calculate the word vector by exerting the related word vector to it with potential function and weighted embedding sampling for simulating the semantic influence between words in same semantic field. We conduct comprehensive experiments and demonstrate that the models using the proposed method can achieve higher accuracy than the baseline defense methods under various adversarial attacks or original testing sets. Moreover, the proposed method is more universal, while it is irrelevant to model structure and will not affect the efficiency of training.
引用
收藏
页码:289 / 301
页数:13
相关论文
共 50 条
  • [1] Defense against adversarial attacks via textual embeddings based on semantic associative field
    Jiacheng Huang
    Long Chen
    Neural Computing and Applications, 2024, 36 : 289 - 301
  • [2] Improving the Semantic Consistency of Textual Adversarial Attacks via Prompt
    Yu, Xiaoyan
    Yin, Qilei
    Shi, Zhixin
    Ma, Yuru
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [3] Deblurring as a Defense against Adversarial Attacks
    Duckworth, William, III
    Liao, Weixian
    Yu, Wei
    2023 IEEE 12TH INTERNATIONAL CONFERENCE ON CLOUD NETWORKING, CLOUDNET, 2023, : 61 - 67
  • [4] Text Adversarial Purification as Defense against Adversarial Attacks
    Li, Linyang
    Song, Demin
    Qiu, Xipeng
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 338 - 350
  • [5] Towards Interpretable Defense Against Adversarial Attacks via Causal Inference
    Min Ren
    Yun-Long Wang
    Zhao-Feng He
    Machine Intelligence Research, 2022, 19 (03) : 209 - 226
  • [6] Towards Interpretable Defense Against Adversarial Attacks via Causal Inference
    Min Ren
    Yun-Long Wang
    Zhao-Feng He
    Machine Intelligence Research, 2022, 19 : 209 - 226
  • [7] Towards Interpretable Defense Against Adversarial Attacks via Causal Inference
    Ren, Min
    Wang, Yun-Long
    He, Zhao-Feng
    MACHINE INTELLIGENCE RESEARCH, 2022, 19 (03) : 209 - 226
  • [8] Defense against adversarial attacks based on color space transformation
    Wang, Haoyu
    Wu, Chunhua
    Zheng, Kangfeng
    NEURAL NETWORKS, 2024, 173
  • [9] AdvRefactor: A Resampling-Based Defense Against Adversarial Attacks
    Jiang, Jianguo
    Li, Boquan
    Yu, Min
    Liu, Chao
    Sun, Jianguo
    Huang, Weiqing
    Lv, Zhiqiang
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2018, PT II, 2018, 11165 : 815 - 825
  • [10] Adversarial Attacks on Node Embeddings via Graph Poisoning
    Bojchevski, Aleksandar
    Guennemann, Stephan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97