Efficient black-box adversarial attacks via alternate query and boundary augmentation

被引:0
作者
Pi, Jiatian [1 ]
Wen, Fusen [2 ]
Xia, Fen [3 ]
Jiang, Ning [3 ]
Wu, Haiying [3 ]
Liu, Qiao [1 ]
机构
[1] Chongqing Normal Univ, Natl Ctr Appl Math, Chongqing, Peoples R China
[2] Chongqing Normal Univ, Sch Comp & Informat Sci, Chongqing, Peoples R China
[3] Mashang Consumer Finance Co Ltd, Chongqing, Peoples R China
基金
中国国家自然科学基金;
关键词
Black-box attack; Query feedback; Transferable prior; Random search;
D O I
10.1016/j.knosys.2025.113604
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most existing query-based black-box attacks use surrogate models as transferable priors to improve query efficiency. However, these methods still suffer from high query times and complexity due to the following three reasons. First, they usually use a transfer-based strategy to find a starting point, which is not conducive to fast optimization. Second, most of them exploit transferable priors in a complex way that severely constrains query efficiency. Third, their performance usually depends on the number of surrogate models and the more surrogate models, the better the performance. To this end, we propose an optimization framework based on fusion attack and boundary augmentation, which make full use of transfer prior and query feedback to achieve a more effective and efficient attack. Specifically, we first use the surrogate model to conduct a warmup attack guided by query feedback, which provides a better starting point for fast optimization. Then, we introduce a data-augmentation-based transferable attack into query-based method for alternative query. Since the alternate attack framework can quickly find out the adversarial area of the target model, it improves the query efficiency. Finally, we design a decision boundary enhancement strategy to make the decision boundary of the model more diverse. This strategy can reduce the number of surrogate models used yet still achieve competitive performance. To validate the effectiveness of the proposed method, we conduct experiments with three victim models on the ImageNet dataset. Extensive experiment results show that our method achieves favorable performance against the state-of-the-art methods. While the proposed method gets a 100% attack success rate, the query times can be reduced by several orders of magnitude.
引用
收藏
页数:12
相关论文
共 46 条
[1]  
[Anonymous], 2016, Found. Trends Optim., DOI DOI 10.1561/2400000013
[2]  
Bhagoji A. N., 2017, arXiv
[3]   Practical Black-Box Attacks on Deep Neural Networks Using Efficient Query Mechanisms [J].
Bhagoji, Arjun Nitin ;
He, Warren ;
Li, Bo ;
Song, Dawn .
COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 :158-174
[4]  
Brendel W, 2018, Arxiv, DOI arXiv:1712.04248
[5]  
Che ZH, 2020, AAAI CONF ARTIF INTE, V34, P3405
[6]  
Chen PY, 2017, PROCEEDINGS OF THE 10TH ACM WORKSHOP ON ARTIFICIAL INTELLIGENCE AND SECURITY, AISEC 2017, P15, DOI 10.1145/3128572.3140448
[7]  
Cheng SY, 2019, ADV NEUR IN, V32
[8]  
Demontis A, 2019, PROCEEDINGS OF THE 28TH USENIX SECURITY SYMPOSIUM, P321
[9]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[10]   Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks [J].
Dong, Yinpeng ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4307-4316