APDL: an adaptive step size method for white-box adversarial attacks

被引:1
作者
Hu, Jiale [1 ,2 ]
Li, Xiang [1 ,2 ]
Liu, Changzheng [1 ,2 ]
Zhang, Ronghua [1 ,2 ]
Tang, Junwei [3 ]
Sun, Yi [1 ,2 ]
Wang, Yuedong [1 ,2 ]
机构
[1] Shihezi Univ, Coll Informat Sci & Technol, Shihezi 832003, Xinjiang, Peoples R China
[2] Xinjiang Prod & Construct Corps, Key Lab Computat Intelligence & Network Informat S, Shihezi, Peoples R China
[3] Wuhan Text Univ, Sch Comp Sci & Artificial Intelligence, Wuhan, Peoples R China
关键词
Adversarial attacks; Deep learning; Image classification; White-box attacks;
D O I
10.1007/s40747-024-01748-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent research has shown that deep learning models are vulnerable to adversarial attacks, including gradient attacks, which can lead to incorrect outputs. The existing gradient attack methods typically rely on repetitive multistep strategies to improve their attack success rates, resulting in longer training times and severe overfitting. To address these issues, we propose an adaptive perturbation-based gradient attack method with dual-loss optimization (APDL). This method adaptively adjusts the single-step perturbation magnitude based on an exponential distance function, thereby accelerating the convergence process. APDL achieves convergence in fewer than 10 iterations, outperforming the traditional nonadaptive methods and achieving a high attack success rate with fewer iterations. Furthermore, to increase the transferability of gradient attacks such as APDL across different models and reduce the effects of overfitting on the training model, we introduce a triple-differential logit fusion (TDLF) method grounded in knowledge distillation principles. This approach mitigates the edge effects associated with gradient attacks by adjusting the hardness and softness of labels. Experiments conducted on ImageNet-compatible datasets demonstrate that APDL is significantly faster than the commonly used nonadaptive methods, whereas the TDLF method exhibits strong transferability.
引用
收藏
页数:12
相关论文
共 42 条
[1]   A New Method for Commercial-Scale Water Purification Selection Using Linguistic Neural Networks [J].
Abdullah, Saleem ;
Almagrabi, Alaa O. ;
Ali, Nawab .
MATHEMATICS, 2023, 11 (13)
[2]   AI-GAN: ATTACK-INSPIRED GENERATION OF ADVERSARIAL EXAMPLES [J].
Bai, Tao ;
Zhao, Jun ;
Zhu, Jinlin ;
Han, Shoudong ;
Chen, Jiefeng ;
Li, Bo ;
Kot, Alex .
2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, :2543-2547
[3]  
Bojarski M, 2016, Arxiv, DOI arXiv:1604.07316
[4]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[5]  
Chen PY, 2017, PROCEEDINGS OF THE 10TH ACM WORKSHOP ON ARTIFICIAL INTELLIGENCE AND SECURITY, AISEC 2017, P15, DOI 10.1145/3128572.3140448
[6]   A Lightweight and Anonymous Mutual Authentication Scheme for Medical Big Data in Distributed Smart Healthcare Systems [J].
Das, Sangjukta ;
Namasudra, Suyel .
IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2024, 21 (04) :1106-1116
[7]   Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks [J].
Dong, Yinpeng ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4307-4316
[8]   Boosting Adversarial Attacks with Momentum [J].
Dong, Yinpeng ;
Liao, Fangzhou ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun ;
Hu, Xiaolin ;
Li, Jianguo .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9185-9193
[9]   Patch-Wise Attack for Fooling Deep Neural Network [J].
Gao, Lianli ;
Zhang, Qilong ;
Song, Jingkuan ;
Liu, Xianglong ;
Shen, Heng Tao .
COMPUTER VISION - ECCV 2020, PT XXVIII, 2020, 12373 :307-322
[10]   Generative Adversarial Networks [J].
Goodfellow, Ian ;
Pouget-Abadie, Jean ;
Mirza, Mehdi ;
Xu, Bing ;
Warde-Farley, David ;
Ozair, Sherjil ;
Courville, Aaron ;
Bengio, Yoshua .
COMMUNICATIONS OF THE ACM, 2020, 63 (11) :139-144