Invisible trigger image: A dynamic neural backdoor attack based on hidden feature

被引:0
作者
Chen, Xiong
Li, Mohan [1 ]
Sun, Yanbin
Tian, Zhihong
机构
[1] Guangzhou Univ, Cyberspace Inst Adv Technol, Guangzhou, Guangdong, Peoples R China
关键词
Neural networks; Backdoor attacks; Security and robustness;
D O I
10.1016/j.neucom.2025.130296
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural backdoor is one of the most significant supply chain threat faced by deep neural networks (DNNs). A common attack method for implanting backdoors in DNNs involves injecting poisoned training data that contains triggers. Triggers are commonly categorized into pixel-level and feature-level triggers. Feature-level triggers typically offer higher stealthiness and robustness compared to pixel-level triggers, but may sometimes be less effective. Additionally, the current generation of feature-level triggers requires unrealistic capabilities from the attacker, such as to control the entire training process or directly manipulate the victim model. To address above challenges, we propose a novel dynamic neural backdoor attack using hidden features. We design a trigger generator with a multi-feature constraint algorithm to overlay hidden features onto clean images in an 'invisible' manner. The trigger generated by our method is dynamic, and the attacker does not need to control the entire training process or directly manipulate the victim model. Experiments show our method outperforms six existing attacks, achieving nearly 100% success with a low poisoning rate of 0.01 while maintaining high accuracy on clean samples. Additionally, it bypasses existing representative backdoor defenses, including three input-level and two model-level defenses.
引用
收藏
页数:12
相关论文
共 38 条
[1]  
Bragg Jonathan, 2021, ADV NEUR IN, V34
[2]  
Chen XY, 2017, Arxiv, DOI arXiv:1712.05526
[3]  
Cheng SY, 2021, AAAI CONF ARTIF INTE, V35, P1148
[4]   SentiNet: Detecting Localized Universal Attacks Against Deep Learning Systems [J].
Chou, Edward ;
Tramer, Florian ;
Pellegrino, Giancarlo .
2020 IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS (SPW 2020), 2020, :48-54
[5]  
Gaikwad SK., 2010, INT J COMPUTER APPL, V10, P16, DOI [DOI 10.5120/1462-1976, 10.5120/1462-1976]
[6]   STRIP: A Defence Against Trojan Attacks on Deep Neural Networks [J].
Gao, Yansong ;
Xu, Change ;
Wang, Derui ;
Chen, Shiping ;
Ranasinghe, Damith C. ;
Nepal, Surya .
35TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE (ACSA), 2019, :113-125
[7]   Image Style Transfer Using Convolutional Neural Networks [J].
Gatys, Leon A. ;
Ecker, Alexander S. ;
Bethge, Matthias .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2414-2423
[8]  
Gu TY, 2019, Arxiv, DOI arXiv:1708.06733
[9]   Few-shot Backdoor Defense Using Shapley Estimation [J].
Guan, Jiyang ;
Tu, Zhuozhuo ;
He, Ran ;
Tao, Dacheng .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :13348-13357
[10]   Towards Inspecting and Eliminating Trojan Backdoors in Deep Neural Networks [J].
Guo, Wenbo ;
Wang, Lun ;
Xu, Yan ;
Xing, Xinyu ;
Du, Min ;
Song, Dawn .
20TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2020), 2020, :162-171