Robust Deep Learning Models against Semantic-Preserving Adversarial Attack

被引:1
|
作者
Zhao, Yunce [1 ,2 ]
Gao, Dashan [1 ,3 ]
Yao, Yinghua [1 ,2 ]
Zhang, Zeqi [4 ]
Mao, Bifei [4 ]
Yao, Xin [1 ]
机构
[1] SUSTech, Dept CSE, Shenzhen, Peoples R China
[2] Univ Technol Sydney, Sydney, NSW, Australia
[3] HKUST, Hong Kong, Peoples R China
[4] Huawei Technol Co Ltd, Shenzhen, Peoples R China
来源
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN | 2023年
基金
中国国家自然科学基金;
关键词
Adversarial Examples; Natural Perturbation; Adversarial Perturbation; Robustness;
D O I
10.1109/IJCNN54540.2023.10191198
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning models can be fooled by small l(p)-norm adversarial perturbations and natural perturbations in terms of attributes. Although the robustness against each perturbation has been explored, it remains a challenge to address the robustness against joint perturbations effectively. In this paper, we study the robustness of deep learning models against joint perturbations by proposing a novel attack mechanism named Semantic-Preserving Adversarial (SPA) attack, which can then be used to enhance adversarial training. Specifically, we introduce an attribute manipulator to generate natural and human-comprehensible perturbations and a noise generator to generate diverse adversarial noises. Based on such combined noises, we optimize both the attribute value and the diversity variable to generate jointlyperturbed samples. For robust training, we adversarially train the deep learning model against the generated joint perturbations. Empirical results on four benchmarks show that the SPA attack causes a larger performance decline with small l1 norm-ball constraints compared to existing approaches. Furthermore, our SPA-enhanced training outperforms existing defense methods against such joint perturbations.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Resisting Deep Learning Models Against Adversarial Attack Transferability via Feature Randomization
    Nowroozi, Ehsan
    Mohammadi, Mohammadreza
    Golmohammadi, Pargol
    Mekdad, Yassine
    Conti, Mauro
    Uluagac, Selcuk
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2024, 17 (01) : 18 - 29
  • [2] Adversarial Defense on Harmony: Reverse Attack for Robust AI Models Against Adversarial Attacks
    Kim, Yebon
    Jung, Jinhyo
    Kim, Hyunjun
    So, Hwisoo
    Ko, Yohan
    Shrivastava, Aviral
    Lee, Kyoungwoo
    Hwang, Uiwon
    IEEE ACCESS, 2024, 12 : 176485 - 176497
  • [3] A Framework for Robust Deep Learning Models Against Adversarial Attacks Based on a Protection Layer Approach
    Al-Andoli, Mohammed Nasser
    Tan, Shing Chiang
    Sim, Kok Swee
    Goh, Pey Yun
    Lim, Chee Peng
    IEEE ACCESS, 2024, 12 : 17522 - 17540
  • [4] Defending Deep Learning Models Against Adversarial Attacks
    Mani, Nag
    Moh, Melody
    Moh, Teng-Sheng
    INTERNATIONAL JOURNAL OF SOFTWARE SCIENCE AND COMPUTATIONAL INTELLIGENCE-IJSSCI, 2021, 13 (01): : 72 - 89
  • [5] Adversarial Attack Against Deep Saliency Models Powered by Non-Redundant Priors
    Che, Zhaohui
    Borji, Ali
    Zhai, Guangtao
    Ling, Suiyi
    Li, Jing
    Tian, Yuan
    Guo, Guodong
    Le Callet, Patrick
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 1973 - 1988
  • [6] Diversity Adversarial Training against Adversarial Attack on Deep Neural Networks
    Kwon, Hyun
    Lee, Jun
    SYMMETRY-BASEL, 2021, 13 (03):
  • [7] Challenging Machine Learning-Based Clone Detectors via Semantic-Preserving Code Transformations
    Zhang, Weiwei
    Guo, Shengjian
    Zhang, Hongyu
    Sui, Yulei
    Xue, Yinxing
    Xu, Yun
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2023, 49 (05) : 3052 - 3070
  • [8] Improving Deep Learning Model Robustness Against Adversarial Attack by Increasing the Network Capacity
    Marchetti, Marco
    Ho, Edmond S. L.
    ADVANCES IN CYBERSECURITY, CYBERCRIMES, AND SMART EMERGING TECHNOLOGIES, 2023, 4 : 85 - 96
  • [9] Multiple-Model Based Defense for Deep Reinforcement Learning Against Adversarial Attack
    Chan, Patrick P. K.
    Wang, Yaxuan
    Kees, Natasha
    Yeung, Daniel S.
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT I, 2021, 12891 : 42 - 53
  • [10] Analyzing the Robustness of Deep Learning Against Adversarial Examples
    Zhao, Jun
    2018 56TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2018, : 1060 - 1064