Learning Patch-Channel Correspondence for Interpretable Face Forgery Detection

被引:18
作者
Hua, Yingying [1 ,2 ]
Shi, Ruixin [1 ,2 ]
Wang, Pengju [1 ,2 ]
Ge, Shiming [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing 100095, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing 100085, Peoples R China
基金
中国国家自然科学基金; 北京市自然科学基金;
关键词
Forgery; Faces; Feature extraction; Deep learning; Decorrelation; Visualization; Task analysis; Face forgery detection; interpretable representation learning; patch-channel correspondence; DEEP NEURAL-NETWORKS;
D O I
10.1109/TIP.2023.3246793
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Beyond high accuracy, good interpretability is very critical to deploy a face forgery detection model for visual content analysis. In this paper, we propose learning patch-channel correspondence to facilitate interpretable face forgery detection. Patch-channel correspondence aims to transform the latent features of a facial image into multi-channel interpretable features where each channel mainly encoders a corresponding facial patch. Towards this end, our approach embeds a feature reorganization layer into a deep neural network and simultaneously optimizes classification task and correspondence task via alternate optimization. The correspondence task accepts multiple zero-padding facial patch images and represents them into channel-aware interpretable representations. The task is solved by step-wisely learning channel-wise decorrelation and patch-channel alignment. Channel-wise decorrelation decouples latent features for class-specific discriminative channels to reduce feature complexity and channel correlation, while patch-channel alignment then models the pairwise correspondence between feature channels and facial patches. In this way, the learned model can automatically discover corresponding salient features associated to potential forgery regions during inference, providing discriminative localization of visualized evidences for face forgery detection while maintaining high detection accuracy. Extensive experiments on popular benchmarks clearly demonstrate the effectiveness of the proposed approach in interpreting face forgery detection without sacrificing accuracy. The source code is available at https://github.com/Jae35/IFFD
引用
收藏
页码:1668 / 1680
页数:13
相关论文
共 74 条
[51]  
Selvaraju RR, 2020, INT J COMPUT VISION, V128, P336, DOI [10.1007/s11263-019-01228-7, 10.1109/ICCV.2017.74]
[52]   Self-Supervised Discovering of Interpretable Features for Reinforcement Learning [J].
Shi, Wenjie ;
Huang, Gao ;
Song, Shiji ;
Wang, Zhuoyuan ;
Lin, Tingyu ;
Wu, Cheng .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (05) :2712-2724
[53]   Detecting Deepfakes with Self-Blended Images [J].
Shiohara, Kaede ;
Yamasaki, Toshihiko .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :18699-18708
[54]  
Smilkov D., 2017, ICML WORKSH VIS DEEP, DOI [10.48550/arXiv.1706.03825, DOI 10.48550/ARXIV.1706.03825]
[55]   Deepfakes Detection Based on Multi Scale Fusion [J].
Sun, Peng ;
Yan, ZhiYuan ;
Shen, Zhe ;
Shi, ShaoPei ;
Dong, Xu .
BIOMETRIC RECOGNITION (CCBR 2021), 2021, 12878 :346-353
[56]   Deferred Neural Rendering: Image Synthesis using Neural Textures [J].
Thies, Justus ;
Zollhofer, Michael ;
Niessner, Matthias .
ACM TRANSACTIONS ON GRAPHICS, 2019, 38 (04)
[57]   Face2Face: Real-time Face Capture and Reenactment of RGB Videos [J].
Thies, Justus ;
Zollhofer, Michael ;
Stamminger, Marc ;
Theobalt, Christian ;
Niessner, Matthias .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2387-2395
[58]   DeepFakes detection across generations: Analysis of facial regions, fusion, and performance evaluation [J].
Tolosana, Ruben ;
Romero-Tapiador, Sergio ;
Vera-Rodriguez, Ruben ;
Gonzalez-Sosa, Ester ;
Fierrez, Julian .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 110
[59]   Deepfakes and beyond: A Survey of face manipulation and fake detection [J].
Tolosana, Ruben ;
Vera-Rodriguez, Ruben ;
Fierrez, Julian ;
Morales, Aythami ;
Ortega-Garcia, Javier .
INFORMATION FUSION, 2020, 64 :131-148
[60]  
Vorontsov E, 2017, PR MACH LEARN RES, V70