Robust Deep Learning Models against Semantic-Preserving Adversarial Attack

被引:1
作者
Zhao, Yunce [1 ,2 ]
Gao, Dashan [1 ,3 ]
Yao, Yinghua [1 ,2 ]
Zhang, Zeqi [4 ]
Mao, Bifei [4 ]
Yao, Xin [1 ]
机构
[1] SUSTech, Dept CSE, Shenzhen, Peoples R China
[2] Univ Technol Sydney, Sydney, NSW, Australia
[3] HKUST, Hong Kong, Peoples R China
[4] Huawei Technol Co Ltd, Shenzhen, Peoples R China
来源
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN | 2023年
基金
中国国家自然科学基金;
关键词
Adversarial Examples; Natural Perturbation; Adversarial Perturbation; Robustness;
D O I
10.1109/IJCNN54540.2023.10191198
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning models can be fooled by small l(p)-norm adversarial perturbations and natural perturbations in terms of attributes. Although the robustness against each perturbation has been explored, it remains a challenge to address the robustness against joint perturbations effectively. In this paper, we study the robustness of deep learning models against joint perturbations by proposing a novel attack mechanism named Semantic-Preserving Adversarial (SPA) attack, which can then be used to enhance adversarial training. Specifically, we introduce an attribute manipulator to generate natural and human-comprehensible perturbations and a noise generator to generate diverse adversarial noises. Based on such combined noises, we optimize both the attribute value and the diversity variable to generate jointlyperturbed samples. For robust training, we adversarially train the deep learning model against the generated joint perturbations. Empirical results on four benchmarks show that the SPA attack causes a larger performance decline with small l1 norm-ball constraints compared to existing approaches. Furthermore, our SPA-enhanced training outperforms existing defense methods against such joint perturbations.
引用
收藏
页数:8
相关论文
共 50 条
[41]   Robust Deep Reinforcement Learning with Adaptive Adversarial Perturbations in Action Space [J].
Liu, Qianmei ;
Kuang, Yufei ;
Wang, Jie .
2024 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN 2024, 2024,
[42]   Toward Robust Discriminative Projections Learning Against Adversarial Patch Attacks [J].
Wang, Zheng ;
Nie, Feiping ;
Wang, Hua ;
Huang, Heng ;
Wang, Fei .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) :18784-18798
[43]   Detecting Adversarial Samples for Deep Learning Models: A Comparative Study [J].
Zhang, Shigeng ;
Chen, Shuxin ;
Liu, Xuan ;
Hua, Chengyao ;
Wang, Weiping ;
Chen, Kai ;
Zhang, Jian ;
Wang, Jianxin .
IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2022, 9 (01) :231-244
[44]   All Points Guided Adversarial Generator for Targeted Attack Against Deep Hashing Retrieval [J].
Tu, Rongxin ;
Kang, Xiangui ;
Tan, Chee Wei ;
Chi, Chi-Hung ;
Lam, Kwok-Yan .
IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2025, 20 :1695-1709
[45]   Adversarial Attacks on Deep Learning Models of Computer Vision: A Survey [J].
Ding, Jia ;
Xu, Zhiwu .
ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT III, 2020, 12454 :396-408
[46]   Deep Adversarial Reinforcement Learning Method to Generate Control Policies Robust Against Worst-Case Value Predictions [J].
Ohashi, Kohei ;
Nakanishi, Kosuke ;
Yasui, Yuji ;
Ishii, Shin .
IEEE ACCESS, 2023, 11 :100798-100809
[47]   QUERY-FREE EMBEDDING ATTACK AGAINST DEEP LEARNING [J].
Liu, Yujia ;
Zhang, Weiming ;
Yu, Nenghai .
2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, :380-386
[48]   Sensitivity based robust learning for stacked autoencoder against evasion attack [J].
Chan, Patrick P. K. ;
Lin, Zhe ;
Hu, Xian ;
Tsang, Eric C. C. ;
Yeung, Daniel S. .
NEUROCOMPUTING, 2017, 267 :572-580
[49]   Feature-Based Adversarial Training for Deep Learning Models Resistant to Transferable Adversarial Examples [J].
Ryu, Gwonsang ;
Choi, Daeseon .
IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (05) :1039-1049
[50]   Multiexpert Adversarial Regularization for Robust and Data-Efficient Deep Supervised Learning [J].
Gholami, Behnam ;
Liu, Qingfeng ;
El-Khamy, Mostafa ;
Lee, Jungwon .
IEEE ACCESS, 2022, 10 :85080-85094