Random transformations to improve mitigation of query-based black-box attacks

被引:1
作者
Ali, Ziad Tariq Muhammad [1 ]
Azad, R. Muhammad Atif [1 ]
Azad, Muhammad Ajmal [1 ]
Holyhead, James [2 ]
Rice, Iain [1 ]
Imran, Ali Shariq [3 ]
机构
[1] Birmingham City Univ, Sch Comp & Digital Technol, Birmingham, England
[2] Covatic, Birmingham, England
[3] Norwegian Univ Sci & Technol, Dept Comp Sci, Trondheim, Norway
关键词
Black-box attacks; Adversarial examples; Randomised defences; Neural networks; SECRET KEY;
D O I
10.1016/j.eswa.2024.125840
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes methods to upstage the best-known defences against query-based black-box attacks. These benchmark defences incorporate gaussian noise into input data during inference to achieve state-of-theart performance in protecting image classification models against the most advanced query-based black-box attacks. Even so there is a need to improve upon them; for example, the widely benchmarked Random noise defense (RND) method has demonstrated limited robustness - achieving only 53.5% and 18.1% with a ResNet50 model on the CIFAR-10 and ImageNet datasets, respectively - against the square attack, which is commonly regarded as the state-of-the-art black-box attack. Therefore, in this work, we propose two alternatives to gaussian noise addition at inference time: random crop-resize and random rotation of the input images. Although these transformations are generally used for data augmentation while training to improve model invariance and generalisation, their protective potential against query-based black-box attacks at inference time is unexplored. Therefore, for the first time, we report that for such well-trained models either of the two transformations can also blunt powerful query-based black-box attacks when used at inference time on three popular datasets. The results show that the proposed randomised transformations outperform RND in terms of robust accuracy against a strong adversary that uses a high budget of 100,000 queries based on expectation over transformation (EOT) of 10, by 0.9% on the CIFAR-10 dataset, 9.4% on the ImageNet dataset and 1.6% on the Tiny ImageNet dataset. Crucially, in two even tougher attack settings, that is, high-confidence adversarial examples and EOT-50 adversary, these transformations are even more effective as the margin of improvement over the benchmarks increases further.
引用
收藏
页数:10
相关论文
共 53 条
[1]   Evaluating Adversarial Robustness of Secret Key-Based Defenses [J].
Ali, Ziad Tariq Muhammad ;
Mohammed, Ameer ;
Ahmad, Imtiaz .
IEEE ACCESS, 2022, 10 :34872-34882
[2]   Square Attack: A Query-Efficient Black-Box Adversarial Attack via Random Search [J].
Andriushchenko, Maksym ;
Croce, Francesco ;
Flammarion, Nicolas ;
Hein, Matthias .
COMPUTER VISION - ECCV 2020, PT XXIII, 2020, 12368 :484-501
[3]  
[Anonymous], 2009, Cifar-10
[4]   Block-Wise Image Transformation With Secret Key for Adversarially Robust Defense [J].
Aprilpyone, Maungmaung ;
Kiya, Hitoshi .
IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 :2709-2723
[5]  
Athalye A, 2018, PR MACH LEARN RES, V80
[6]   Integrating Convolutional Neural Network and Multiresolution Segmentation for Land Cover and Land Use Mapping Using Satellite Imagery [J].
Atik, Saziye Ozge ;
Ipbuker, Cengizhan .
APPLIED SCIENCES-BASEL, 2021, 11 (12)
[7]  
Biggio Battista, 2013, Machine Learning and Knowledge Discovery in Databases. European Conference, ECML PKDD 2013. Proceedings: LNCS 8190, P387, DOI 10.1007/978-3-642-40994-3_25
[8]   On the Effectiveness of Small Input Noise for Defending Against Query-based Black-Box Attacks [J].
Byun, Junyoung ;
Go, Hyojun ;
Kim, Changick .
2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, :3819-3828
[9]  
Carlini N, 2019, Arxiv, DOI arXiv:1902.06705
[10]   Cryptanalytic Extraction of Neural Network Models [J].
Carlini, Nicholas ;
Jagielski, Matthew ;
Mironov, Ilya .
ADVANCES IN CRYPTOLOGY - CRYPTO 2020, PT III, 2020, 12172 :189-218