Black box phase-based adversarial attacks on image classifiers

被引:0
作者
Hodes, Scott G. [1 ,2 ]
Blose, Kory J. [1 ,3 ]
Kane, Timothy J. [1 ,2 ]
机构
[1] Penn State Univ, Appl Res Lab, POB 30, State Coll, PA 16804 USA
[2] Penn State Univ, Sch Elect Engn & Comp Sci, 121 Elect Engn East Bldg, University Pk, PA 16802 USA
[3] Penn State Univ, Dept Agr & Biol Engn, 105 Agr Engn Bldg, University Pk, PA 16802 USA
来源
AUTOMATIC TARGET RECOGNITION XXXIV | 2024年 / 13039卷
关键词
adversarial attack; black box attack; Fourier optics; image classifier; neural networks; spatial light modulator;
D O I
10.1117/12.3013308
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
We propose a new method of utilizing a spatial light modulator to generate adversarial examples against image classifiers within a black box scenario. The method incorporates a simple-shape-focused strategy that queries the target network and estimates the effect of perturbing specific regions of the Fourier plane. This work is an extension of previous work that uses a spatial light modulator to perturb the phase of incoming light to generate adversarial patterns using l(2)-norm optimization. Our new method simply uses the final logits of the target network, allowing for it to be used not only in "white box" scenarios but also in the information-constrained "black box" scenarios. Our shape-based algorithm is shown to be widely effective on the original dataset benchmark without the requirement of knowledge about the target network architecture. Our experiments explore how manipulating the size, shape, number, and magnitude of the regions tested affects the efficacy and pattern cycles needed to generate a successful attack. Different combinations showed a range of average efficacy between 32% and 63% under a consistent objective function. Our new method also proved to be effective on a smaller dataset (meaning fewer classes for classification to be misdirected towards). We validate our method using a physical setup.
引用
收藏
页数:20
相关论文
共 19 条
[1]   GenAttack: Practical Black-box Attacks with Gradient-Free Optimization [J].
Alzantot, Moustafa ;
Sharma, Yash ;
Chakraborty, Supriyo ;
Zhang, Huan ;
Hsieh, Cho-Jui ;
Srivastava, Mani B. .
PROCEEDINGS OF THE 2019 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE (GECCO'19), 2019, :1111-1119
[2]  
Brendel W., 2017, Decision-based adversarial attacks: Reliable attacks against black-box machine learning models, V12
[3]  
Demontis A, 2019, PROCEEDINGS OF THE 28TH USENIX SECURITY SYMPOSIUM, P321
[4]  
Goodman J. W., 2005, INTRO FOURIER OPTICS
[5]  
Houben S, 2013, IEEE IJCNN
[6]  
Ilyas Andrew, 2018, P MACHINE LEARNING R, V80
[7]   Engineering pupil function for optical adversarial attacks [J].
Kim, Kyulim ;
Kim, Jeongsoo ;
Song, Seungri ;
Choi, Jun-Ho ;
Joo, Chulmin ;
Lee, Jong-Seok .
OPTICS EXPRESS, 2022, 30 (05) :6500-6518
[8]  
Kurakin A, 2018, SPRING SER CHALLENGE, P195, DOI 10.1007/978-3-319-94042-7_11
[9]   OPA2D: One-Pixel Attack, Detection, and Defense in Deep Neural Networks [J].
Nguyen-Son, Hoang-Quoc ;
Thao, Tran Phuong ;
Hidano, Seira ;
Bracamonte, Vanessa ;
Kiyomoto, Shinsaku ;
Yamaguchi, Rie Shigetomi .
2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
[10]   Practical Black-Box Attacks against Machine Learning [J].
Papernot, Nicolas ;
McDaniel, Patrick ;
Goodfellow, Ian ;
Jha, Somesh ;
Celik, Z. Berkay ;
Swami, Ananthram .
PROCEEDINGS OF THE 2017 ACM ASIA CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (ASIA CCS'17), 2017, :506-519