HyGloadAttack: Hard-label black-box textual adversarial attacks via hybrid optimization

被引:20
|
作者
Liu, Zhaorong [1 ,2 ,4 ]
Xiong, Xi [1 ,2 ,4 ]
Li, Yuanyuan [3 ]
Yu, Yan [1 ,2 ,4 ]
Lu, Jiazhong [1 ,2 ,4 ]
Zhang, Shuai [5 ]
Xiong, Fei [6 ]
机构
[1] Chengdu Univ Informat Technol, Sch Cybersecur, Chengdu 610225, Peoples R China
[2] Adv Cryptog & Syst Secur Key Lab Sichuan Prov, Chengdu 610225, Peoples R China
[3] Sichuan Univ, West China Sch Med, Chengdu 610041, Peoples R China
[4] SUGON Ind Control & Secur Ctr, Chengdu 610225, Peoples R China
[5] Beijing Inst Technol, Informat Syst & Secur & Countermeasures Expt Ctr, Beijing 100081, Peoples R China
[6] Beijing Jiaotong Univ, Key Lab Commun & Informat Syst, Beijing Municipal Commiss Educ, Beijing 100044, Peoples R China
基金
中国国家自然科学基金;
关键词
Adversarial attack; Robustness; Black-box; Hard-label;
D O I
10.1016/j.neunet.2024.106461
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hard -label black -box textual adversarial attacks present a highly challenging task due to the discrete and nondifferentiable nature of text data and the lack of direct access to the model's predictions. Research in this issue is still in its early stages, and the performance and efficiency of existing methods has potential for improvement. For instance, exchange -based and gradient -based attacks may become trapped in local optima and require excessive queries, hindering the generation of adversarial examples with high semantic similarity and low perturbation under limited query conditions. To address these issues, we propose a novel framework called HyGloadAttack ( ad versarial Attack s via Hy brid optimization and Glo bal random initialization) for crafting high -quality adversarial examples. HyGloadAttack utilizes a perturbation matrix in the word embedding space to find nearby adversarial examples after global initialization and selects synonyms that maximize similarity while maintaining adversarial properties. Furthermore, we introduce a gradient -based quick search method to accelerate the search process of optimization. Extensive experiments on five datasets of text classification and natural language inference, as well as two real APIs, demonstrate the significant superiority of our proposed HyGloadAttack method over state-of-the-art baseline methods.
引用
收藏
页数:15
相关论文
共 44 条
  • [1] DFDS: Data-Free Dual Substitutes Hard-Label Black-Box Adversarial Attack
    Jiang, Shuliang
    He, Yusheng
    Zhang, Rui
    Kang, Zi
    Xia, Hui
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT III, KSEM 2024, 2024, 14886 : 274 - 285
  • [2] Simple and Efficient Hard Label Black-box Adversarial Attacks in Low Query Budget Regimes
    Shukla, Satya Narayan
    Sahu, Anit Kumar
    Willmott, Devin
    Kolter, Zico
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 1461 - 1469
  • [3] Semantic-Aware Adaptive Binary Search for Hard-Label Black-Box Attack
    Ma, Yiqing
    Lucke, Kyle
    Xian, Min
    Vakanski, Aleksandar
    COMPUTERS, 2024, 13 (08)
  • [4] Automatic Selection Attacks Framework for Hard Label Black-Box Models
    Liu, Xiaolei
    Li, Xiaoyu
    Zheng, Desheng
    Bai, Jiayu
    Peng, Yu
    Zhang, Shibin
    IEEE INFOCOM 2022 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (INFOCOM WKSHPS), 2022,
  • [5] Black-Box Dissector: Towards Erasing-Based Hard-Label Model Stealing Attack
    Wang, Yixu
    Li, Jie
    Liu, Hong
    Wang, Yan
    Wu, Yongjian
    Huang, Feiyue
    Ji, Rongrong
    COMPUTER VISION - ECCV 2022, PT V, 2022, 13665 : 192 - 208
  • [6] Black-box attacks on dynamic graphs via adversarial topology perturbations
    Tao, Haicheng
    Cao, Jie
    Chen, Lei
    Sun, Hongliang
    Shi, Yong
    Zhu, Xingquan
    NEURAL NETWORKS, 2024, 171 : 308 - 319
  • [7] Efficient text-based evolution algorithm to hard-label adversarial attacks on text
    Peng, Hao
    Wang, Zhe
    Zhao, Dandan
    Wu, Yiming
    Han, Jianming
    Guo, Shixin
    Ji, Shouling
    Zhong, Ming
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2023, 35 (05)
  • [8] Physical Black-Box Adversarial Attacks Through Transformations
    Jiang, Wenbo
    Li, Hongwei
    Xu, Guowen
    Zhang, Tianwei
    Lu, Rongxing
    IEEE TRANSACTIONS ON BIG DATA, 2023, 9 (03) : 964 - 974
  • [9] Black-box adversarial attacks by manipulating image attributes
    Wei, Xingxing
    Guo, Ying
    Li, Bo
    INFORMATION SCIENCES, 2021, 550 : 285 - 296
  • [10] WordBlitz: An Efficient Hard-Label Textual Adversarial Attack Method Jointly Leveraging Adversarial Transferability and Word Importance
    Li, Xiangge
    Luo, Hong
    Sun, Yan
    APPLIED SCIENCES-BASEL, 2024, 14 (09):