Generating adversarial samples by manipulating image features with auto-encoder

被引:0
作者
Yang, Jianxin [1 ]
Shao, Mingwen [1 ]
Liu, Huan [1 ]
Zhuang, Xinkai [1 ]
机构
[1] China Univ Petr East China, Coll Comp Sci & Technol, Qingdao 266000, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep neural networks; Adversarial attacks; Adversarial samples; Style features; ATTACK;
D O I
10.1007/s13042-023-01778-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing adversarial attack methods usually add perturbations directly to the pixel space of an image, resulting in significant local noise in the image. Besides, the performance of existing attack methods is affected by various pixel-space based defense strategies. In this paper, we propose a novel method to generate adversarial examples by adding perturbations to the feature space. Specifically, the perturbation of the feature space is induced by a style-shifting-based network architecture called AdvAdaIN. Furthermore, we expose the feature space to the attacker via an encoder, and then the perturbation is injected into the feature space by AdvAdaIN. Simultaneously, due to the specificity of feature space perturbations, we trained a decoder to reflect the changes in feature space to pixel space and ensure that the perturbations are not easily detected. Meanwhile, we align the original image with another image in the feature space, adding additional adversarial information to the model. In addition, we can generate diverse adversarial samples by varying the perturbation parameters, which mainly change the overall color and brightness of the image. Experiments demonstrate that the proposed method outperforms existing methods and produces more natural adversarial samples when facing defensive strategies.
引用
收藏
页码:2499 / 2509
页数:11
相关论文
共 39 条
  • [1] Advances in Adversarial Attacks and Defenses in Computer Vision: A Survey
    Akhtar, Naveed
    Mian, Ajmal
    Kardan, Navid
    Shah, Mubarak
    [J]. IEEE ACCESS, 2021, 9 : 155161 - 155196
  • [2] Bhattad A., 2020, 8 INT C LEARNING REP
  • [3] Towards Evaluating the Robustness of Neural Networks
    Carlini, Nicholas
    Wagner, David
    [J]. 2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, : 39 - 57
  • [4] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [5] Boosting Adversarial Attacks with Momentum
    Dong, Yinpeng
    Liao, Fangzhou
    Pang, Tianyu
    Su, Hang
    Zhu, Jun
    Hu, Xiaolin
    Li, Jianguo
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 9185 - 9193
  • [6] Dumoulin V., 2017, 5 INT C LEARN REPR I
  • [7] Robust Physical-World Attacks on Deep Learning Visual Classification
    Eykholt, Kevin
    Evtimov, Ivan
    Fernandes, Earlence
    Li, Bo
    Rahmati, Amir
    Xiao, Chaowei
    Prakash, Atul
    Kohno, Tadayoshi
    Song, Dawn
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 1625 - 1634
  • [8] Gatys LA., 2015, A neural algorithm of artistic style, V16, P326, DOI DOI 10.1167/16.12.326
  • [9] Goodfellow I., 2015, P INT C LEARN REPR
  • [10] Gu Tianyu, 2017, CoRR abs/1708.06733