Crafting imperceptible and transferable adversarial examples: leveraging conditional residual generator and wavelet transforms to deceive deepfake detection

被引:0
作者
Li, Zhiyuan [1 ]
Jin, Xin [1 ]
Jiang, Qian [1 ]
Wang, Puming [1 ]
Lee, Shin-Jye [2 ]
Yao, Shaowen [1 ]
Zhou, Wei [1 ]
机构
[1] Yunnan Univ, Kunming, Yunnan, Peoples R China
[2] Natl Yang Ming Chiao Tung Univ, Hsinchu, Taiwan
基金
中国国家自然科学基金;
关键词
Deepfake detection; Adversarial examples; Imperceptible; Transferability; Black-box attacks;
D O I
10.1007/s00371-024-03605-x
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
The malicious abuse of deepfakes has raised serious ethical, security, and privacy concerns, eroding public trust in digital media. While existing deepfake detectors can detect fake images, they are vulnerable to adversarial attacks. Although various adversarial attacks have been explored, most are white-box attacks difficult to realize in practice, and the generated adversarial examples have poor quality easily noticeable to the human eye. For this detection task, the goal should be to generate adversarial examples that can deceive detectors while maintaining high quality and authenticity. We propose a method to generate imperceptible and transferable adversarial examples aimed at fooling unknown deepfake detectors. The method combines a conditional residual generator with an accessible detector as a surrogate model, utilizing the detector's relative distance loss function to generate highly transferable adversarial examples. Discrete wavelet transform is also introduced to enhance image quality. Extensive experiments demonstrate that the adversarial examples generated by our method not only possess excellent visual quality but also effectively deceive various detectors, exhibiting superior cross-detector transferability in black-box attacks. Our code is available at:https://github.com/SiSuiyuHang/ITA.
引用
收藏
页码:3329 / 3344
页数:16
相关论文
共 62 条
[1]  
Afchar D, 2018, IEEE INT WORKS INFOR
[2]  
Bingen Li H.Z., 2021, UNSUPERVISEDMONOCULA
[3]   Evading Deepfake-Image Detectors with White- and Black-Box Attacks [J].
Carlini, Nicholas ;
Farid, Hany .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, :2804-2813
[4]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[5]  
Chen PY, 2017, PROCEEDINGS OF THE 10TH ACM WORKSHOP ON ARTIFICIAL INTELLIGENCE AND SECURITY, AISEC 2017, P15, DOI 10.1145/3128572.3140448
[6]   Deep Colorization [J].
Cheng, Zezhou ;
Yang, Qingxiong ;
Sheng, Bin .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :415-423
[7]   Xception: Deep Learning with Depthwise Separable Convolutions [J].
Chollet, Francois .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1800-1807
[8]   Saliency Attack: Towards Imperceptible Black-box Adversarial Attack [J].
Dai, Zeyu ;
Liu, Shengcai ;
Li, Qing ;
Tang, Ke .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2023, 14 (03)
[9]   Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks [J].
Dong, Yinpeng ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4307-4316
[10]   Boosting Adversarial Attacks with Momentum [J].
Dong, Yinpeng ;
Liao, Fangzhou ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun ;
Hu, Xiaolin ;
Li, Jianguo .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9185-9193