Undercover Deepfakes: Detecting Fake Segments in Videos

被引:2
作者
Saha, Sanjay [1 ]
Perera, Rashindrie [2 ]
Seneviratne, Sachith [2 ]
Malepathirana, Tamasha [2 ]
Rasnayaka, Sanka [1 ]
Geethika, Deshani [2 ]
Sim, Terence [1 ]
Halgamuge, Saman [2 ]
机构
[1] Natl Univ Singapore, Singapore, Singapore
[2] Univ Melbourne, Melbourne, Vic, Australia
来源
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW | 2023年
关键词
D O I
10.1109/ICCVW60793.2023.00048
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The recent renaissance in generative models, driven primarily by the advent of diffusion models and iterative improvement in GAN methods, has enabled many creative applications. However, each advancement is also accompanied by a rise in the potential for misuse. In the arena of the deepfake generation, this is a key societal issue. In particular, the ability to modify segments of videos using such generative techniques creates a new paradigm of deepfakes which are mostly real videos altered slightly to distort the truth. This paradigm has been under-explored by the current deepfake detection methods in the academic literature. In this paper, we present a deepfake detection method that can address this issue by performing deepfake prediction at the frame and video levels. To facilitate testing our method, we prepared a new benchmark dataset where videos have both real and fake frame sequences with very subtle transitions. We provide a benchmark on the proposed dataset with our detection method which utilizes the Vision Transformer based on Scaling and Shifting [38] to learn spatial features, and a Timeseries Transformer to learn temporal features of the videos to help facilitate the interpretation of possible deepfakes. Extensive experiments on a variety of deepfake generation methods show excellent results by the proposed method on temporal segmentation and classical video-level predictions as well. In particular, the paradigm we address will form a powerful tool for the moderation of deepfakes, where human oversight can be better targeted to the parts of videos suspected of being deepfakes. All experiments can be reproduced at: github.com/rgb91/temporal-deepfake-segmentation.
引用
收藏
页码:415 / 425
页数:11
相关论文
共 73 条
[1]  
Passos LA, 2022, Arxiv, DOI arXiv:2202.06095
[2]  
Afchar D, 2018, IEEE INT WORKS INFOR
[3]  
Ahmed SR, 2022, 2022 INT C HUMAN COM, P1
[4]   Deepfake Video Detection through Optical Flow based CNN [J].
Amerini, Irene ;
Galteri, Leonardo ;
Caldelli, Roberto ;
Del Bimbo, Alberto .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, :1205-1207
[5]  
Amoroso R, 2024, Arxiv, DOI arXiv:2304.00500
[6]  
Bayar B., 2016, ACM WORKSH INF HID M, P5, DOI 10.1145/2909827.2930786
[7]  
Bubeck S, 2023, Arxiv, DOI [arXiv:2303.12712, DOI 10.48550/ARXIV.2303.12712]
[8]  
Cai Zhixi, 2022, 2022 INT C DIGITAL I, P1
[9]   Self-supervised Learning of Adversarial Example: Towards Good Generalizations for Deepfake Detection [J].
Chen, Liang ;
Zhang, Yong ;
Song, Yibing ;
Liu, Lingqiao ;
Wang, Jue .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :18689-18698
[10]  
Chen SF, 2022, Arxiv, DOI arXiv:2205.13535