Exploring Multi-View Pixel Contrast for General and Robust Image Forgery Localization

被引:8
作者
Lou, Zijie [1 ,2 ]
Cao, Gang [1 ,2 ]
Guo, Kun [1 ,2 ]
Yu, Lifang [3 ]
Weng, Shaowei [4 ]
机构
[1] Commun Univ China, Sch Comp & Cyber Sci, Beijing 100024, Peoples R China
[2] Commun Univ China, State Key Lab Media Convergence & Commun, Beijing, Peoples R China
[3] Beijing Inst Grap Commun, Dept Informat Engn, Beijing 100026, Peoples R China
[4] Fujian Univ Technol, Fujian Prov Key Lab Big Data Min & Applicat, Fuzhou 350118, Peoples R China
基金
中国国家自然科学基金;
关键词
Location awareness; Forgery; Training; Head; Contrastive learning; Feature extraction; Protocols; Forensics; Testing; Accuracy; Digital forensics; image forensics; image forgery localization; multi-view contrastive learning; large scale testing;
D O I
10.1109/TIFS.2025.3541957
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Image forgery localization, which aims to segment tampered regions in an image, is a fundamental yet challenging digital forensic task. While some deep learning-based forensic methods have achieved impressive results, they directly learn pixel-to-label mappings without fully exploiting the relationship between pixels in the feature space. To address such deficiency, we propose a Multi-view Pixel-wise Contrastive algorithm (MPC) for image forgery localization. Specifically, we first pre-train the feature extraction backbone network with a supervised contrastive loss to model pixel relationships in view of within-image, cross-scale and cross-modality. That is aimed at increasing intra-class compactness and inter-class separability. Then the localization head is fine-tuned using cross-entropy loss, resulting in a better forged pixel localizer. The MPC is trained on three different scale training datasets to make a comprehensive and fair comparison with existing image forgery localization algorithms. Extensive test results on over ten public datasets show that the proposed MPC achieves higher generalization performance and robustness than the state-of-the-arts. It is particularly noteworthy that our approach maintains a high level of localization accuracy under various post-processing combinations that approximate real-world scenarios, as well as when confronted with novel intelligent editing techniques. Finally, comprehensive and detailed ablation experiments demonstrate the reasonableness of MPC.
引用
收藏
页码:2329 / 2341
页数:13
相关论文
共 65 条
[1]  
[Anonymous], 2013, IEEE IFS-TC image forensics challenge dataset
[2]   Hybrid LSTM and Encoder-Decoder Architecture for Detection of Image Forgeries [J].
Bappy, Jawadul H. ;
Simons, Cody ;
Nataraj, Lakshmanan ;
Manjunath, B. S. ;
Roy-Chowdhury, Amit K. .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (07) :3286-3300
[3]   RRU-Net: The Ringed Residual U-Net for Image Splicing Forgery Detection [J].
Bi, Xiuli ;
Wei, Yang ;
Xiao, Bin ;
Li, Weisheng .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, :30-39
[4]   Illuminant-Based Transformed Spaces for Image Forensics [J].
Carvalho, Tiago ;
Faria, Fabio A. ;
Pedrini, Helio ;
Torres, Ricardo da S. ;
Rocha, Anderson .
IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2016, 11 (04) :720-733
[5]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[6]  
Chen T, 2020, PR MACH LEARN RES, V119
[7]   Noiseprint: A CNN-Based Camera Model Fingerprint [J].
Cozzolino, Davide ;
Verdoliva, Luisa .
IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2020, 15 (01) :144-159
[8]   MVSS-Net: Multi-View Multi-Scale Supervised Networks for Image Manipulation Detection [J].
Dong, Chengbo ;
Chen, Xinru ;
Hu, Ruohan ;
Cao, Juan ;
Li, Xirong .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) :3539-3553
[9]  
Dosovitskiy A., 2021, P INT C LEARN REPR, DOI [10.48550/arXiv.2010.11929, DOI 10.48550/ARXIV.2010.11929]
[10]  
Fan Y, 2015, IEEE IMAGE PROC, P2940, DOI 10.1109/ICIP.2015.7351341