Detecting Compressed Deepfake Videos in Social Networks Using Frame-Temporality Two-Stream Convolutional Network

被引:128
作者
Hu, Juan [1 ]
Liao, Xin [1 ,2 ]
Wang, Wei [3 ]
Qin, Zheng [1 ]
机构
[1] Hunan Univ, Coll Comp Sci & Elect Engn, Changsha 410082, Peoples R China
[2] Chinese Acad Sci, Inst Informat Engn, State Key Lab Informat Secur, Beijing 100093, Peoples R China
[3] Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing 100190, Peoples R China
基金
中国国家自然科学基金;
关键词
Videos; Information integrity; Feature extraction; Streaming media; Faces; Forensics; Social networking (online); Video forensics; compressed Deepfake videos; frame-level stream; temporality-level stream; FORENSICS;
D O I
10.1109/TCSVT.2021.3074259
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The development of technologies that can generate Deepfake videos is expanding rapidly. These videos are easily synthesized without leaving obvious traces of manipulation. Though forensically detection in high-definition video datasets has achieved remarkable results, the forensics of compressed videos is worth further exploring. In fact, compressed videos are common in social networks, such as videos from Instagram, Wechat, and Tiktok. Therefore, how to identify compressed Deepfake videos becomes a fundamental issue. In this paper, we propose a two-stream method by analyzing the frame-level and temporality-level of compressed Deepfake videos. Since the video compression brings lots of redundant information to frames, the proposed frame-level stream gradually prunes the network to prevent the model from fitting the compression noise. Aiming at the problem that the temporal consistency in Deepfake videos might be ignored, we apply a temporality-level stream to extract temporal correlation features. When combined with scores from the two streams, our proposed method performs better than the state-of-the-art methods in compressed Deepfake videos detection.
引用
收藏
页码:1089 / 1102
页数:14
相关论文
共 39 条
[1]  
Afchar D, 2018, IEEE INT WORKS INFOR
[2]   Sequential and Patch Analyses for Object Removal Video Forgery Detection and Localization [J].
Aloraini, Mohammed ;
Sharifzadeh, Mehdi ;
Schonfeld, Dan .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (03) :917-930
[3]  
[Anonymous], 2018, FACESWAP GITHUB NONO
[4]  
[Anonymous], 2018, DeepFakes Github-Non Official Project Based on Original DeepFakes Thread
[5]  
Baroncini V., 2009, 2009 17th European Signal Processing Conference (EUSIPCO 2009), P564
[6]   Deep Neural Networks for No-Reference and Full-Reference Image Quality Assessment [J].
Bosse, Sebastian ;
Maniry, Dominique ;
Mueller, Klaus-Robert ;
Wiegand, Thomas ;
Samek, Wojciech .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (01) :206-219
[7]   Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset [J].
Carreira, Joao ;
Zisserman, Andrew .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :4724-4733
[8]   Automatic Detection of Object-Based Forgery in Advanced Video [J].
Chen, Shengda ;
Tan, Shunquan ;
Li, Bin ;
Huang, Jiwu .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2016, 26 (11) :2138-2151
[9]  
Cheng Y., SOFTWARE ENG KNOWLED, V115
[10]   A PatchMatch-Based Dense-Field Algorithm for Video Copy-Move Detection and Localization [J].
D'Amiano, Luca ;
Cozzolino, Davide ;
Poggi, Giovanni ;
Verdoliva, Luisa .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (03) :669-682