Generalizable face forgery detection based on adaptive spatial-frequency information mining

被引:0
作者
Qi, Yongfeng [1 ]
Xie, Hongli [1 ,2 ]
Gao, Yajuan [1 ]
Lin, Yuanzhe [1 ]
Zhang, Heng [1 ]
Han, Haixi [1 ]
机构
[1] Northwest Normal Univ, Coll Comp Sci & Engn, Lanzhou 730070, Gansu, Peoples R China
[2] Northwest Normal Univ, Key Lab Cryptog & Data Analyt, Lanzhou 730070, Gansu, Peoples R China
基金
中国国家自然科学基金;
关键词
Deepfake detection; Face forgery detection; Frequency-aware learning; Spatial texture enhancement; DCT;
D O I
10.1007/s00530-025-01893-4
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the current field of face forgery detection, researchers are focused on recognizing forgery cues through the combination of frequency information and convolutional neural networks (CNN). However, existing methods often fail to capture spatial correlations with image content when extracting frequency features, making it difficult to accurately recognize highly simulated forged images. In addition, these methods perform well on homogeneous datasets, but their effectiveness decreases significantly when evaluated on cross-dataset samples. To address these issues, we propose a novel adaptive spatial-frequency information mining (ASFIM) method for generalizable face forgery detection. Specifically, the ASFIM method first processes the original RGB image through a frequency-aware learning module. This module extracts forgery frequency information closely related to the image content, which is subsequently used as input for frequency branching. Next, a spatial texture enhancement module is introduced to enable interaction between spatial and frequency features at an early stage. This approach not only strengthens the expressiveness of forgery features in the spatial domain but also provides an effective guide for recognizing forgery cues in the frequency domain. Finally, we designed the cross-domain interactive attention (CDIA) module to enhance forgery cues by deeply fusing spatial texture and frequency-aware features. Extensive experimental results demonstrate that the proposed ASFIM method outperforms various advanced methods in terms of generalization ability across challenging benchmark tests.
引用
收藏
页数:16
相关论文
共 60 条
[1]   DISCRETE COSINE TRANSFORM [J].
AHMED, N ;
NATARAJAN, T ;
RAO, KR .
IEEE TRANSACTIONS ON COMPUTERS, 1974, C 23 (01) :90-93
[2]   Self-supervised Learning of Adversarial Example: Towards Good Generalizations for Deepfake Detection [J].
Chen, Liang ;
Zhang, Yong ;
Song, Yibing ;
Liu, Lingqiao ;
Wang, Jue .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :18689-18698
[3]  
Chen S, 2021, AAAI CONF ARTIF INTE, V35, P1081
[4]  
Chen ZH, 2020, Arxiv, DOI arXiv:2005.02958
[5]   Xception: Deep Learning with Depthwise Separable Convolutions [J].
Chollet, Francois .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1800-1807
[6]  
Cozzolino D, 2019, Arxiv, DOI [arXiv:1812.02510, 10.48550/arXiv.1812.02510]
[7]   On the Detection of Digital Face Manipulation [J].
Dang, Hao ;
Liu, Feng ;
Stehouwer, Joel ;
Liu, Xiaoming ;
Jain, Anil K. .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :5780-5789
[8]  
Dolhansky B, 2019, Arxiv, DOI arXiv:1910.08854
[9]   Implicit Identity Leakage: The Stumbling Block to Improving Deepfake Detection Generalization [J].
Dong, Shichao ;
Wang, Jin ;
Ji, Renhe ;
Liang, Jiajun ;
Fan, Haoqing ;
He, Zheng .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, :3994-4004
[10]   CTAFFNet: CNN-Transformer Adaptive Feature Fusion Object Detection Algorithm for Complex Traffic Scenarios [J].
Dong, Xinlong ;
Shi, Peicheng ;
Liang, Taonian ;
Yang, Aixi .
TRANSPORTATION RESEARCH RECORD, 2025, 2679 (01) :1947-1965