Evading DeepFake Detectors via Adversarial Statistical Consistency

被引:22
作者
Hou, Yang [1 ]
Guo, Qing [2 ]
Huang, Yihao [3 ]
Xie, Xiaofei [4 ]
Ma, Lei [5 ,6 ]
Zhao, Jianjun [1 ]
机构
[1] Kyushu Univ, Fukuoka, Japan
[2] ASTAR, Ctr Frontier Res CFAR, Singapore, Singapore
[3] Nanyang Technol Univ, Singapore, Singapore
[4] Singapore Management Univ, Singapore, Singapore
[5] Univ Alberta, Edmonton, AB, Canada
[6] Univ Tokyo, Tokyo, Japan
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2023年
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
10.1109/CVPR52729.2023.01181
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, as various realistic face forgery techniques known as DeepFake improves by leaps and bounds, more and more DeepFake detection techniques have been proposed. These methods typically rely on detecting statistical differences between natural (i.e., real) and DeepFake-generated images in both spatial and frequency domains. In this work, we propose to explicitly minimize the statistical differences to evade state-of-the-art DeepFake detectors. To this end, we propose a statistical consistency attack (StatAttack) against DeepFake detectors, which contains two main parts. First, we select several statistical-sensitive natural degradations (i.e., exposure, blur, and noise) and add them to the fake images in an adversarial way. Second, we find that the statistical differences between natural and DeepFake images are positively associated with the distribution shifting between the two kinds of images, and we propose to use a distribution-aware loss to guide the optimization of different degradations. As a result, the feature distributions of generated adversarial examples is close to the natural images. Furthermore, we extend the StatAttack to a more powerful version, MStatAttack, where we extend the single-layer degradation to multi-layer degradations sequentially and use the loss to tune the combination weights jointly. Comprehensive experimental results on four spatial-based detectors and two frequency-based detectors with four datasets demonstrate the effectiveness of our proposed attack method in both white-box and black-box settings.
引用
收藏
页码:12271 / 12280
页数:10
相关论文
共 53 条
[1]   CNN Detection of GAN-Generated Face Images based on Cross-Band Co-occurrences Analysis [J].
Barni, Mauro ;
Kailas, Kassem ;
Nowroozi, Ehsan ;
Tondi, Benedetta .
2020 IEEE INTERNATIONAL WORKSHOP ON INFORMATION FORENSICS AND SECURITY (WIFS), 2020,
[2]   Evading Deepfake-Image Detectors with White- and Black-Box Attacks [J].
Carlini, Nicholas ;
Farid, Hany .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, :2804-2813
[3]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[4]  
Cheng SY, 2019, ADV NEUR IN, V32
[5]   StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation [J].
Choi, Yunjey ;
Choi, Minje ;
Kim, Munyoung ;
Ha, Jung-Woo ;
Kim, Sunghun ;
Choo, Jaegul .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8789-8797
[6]  
Cozzolino Davide, 2018, CORR
[7]  
Deepfake, DEEPFAKE FACESWAP
[8]   Boosting Adversarial Attacks with Momentum [J].
Dong, Yinpeng ;
Liao, Fangzhou ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun ;
Hu, Xiaolin ;
Li, Jianguo .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9185-9193
[9]   Watch your Up-Convolution: CNN Based Generative Deep Neural Networks are Failing to Reproduce Spectral Distributions [J].
Durall, Ricard ;
Keuper, Margret ;
Keuper, Janis .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :7887-7896
[10]  
Dziugaite GK, 2015, UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, P258