Exploiting Complementary Dynamic Incoherence for DeepFake Video Detection

被引:22
|
作者
Wang, Hanyi [1 ]
Liu, Zihan [1 ]
Wang, Shilin [1 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Elect Informat & Elect Engn, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
DeepFake video detection; video forensics;
D O I
10.1109/TCSVT.2023.3238517
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Recently, manipulated videos based on DeepFake technology have spread widely on social media, causing concerns about the authenticity of video content and personal privacy protection. Although existing DeepFake detection methods achieve remarkable progress in some specific scenarios, their detection performance usually drops drastically when detecting unseen manipulation methods. Compared with static information such as human face, dynamic information depicting the movements of facial features is more difficult to forge without leaving visual or statistical traces. Hence, in order to achieve better generalization ability, we focus on dynamic information analysis to disclose such traces and propose a novel Complementary Dynamic Interaction Network (CDIN). Inspired by the DeepFake detection methods based on mouth region analysis, both the global (entire face) and local (mouth region) dynamics are analyzed with properly designed network branches, respectively, and their feature maps at various levels are communicated with each other using a newly proposed Complementary Cross Dynamics Fusion Module (CCDFM). With CCDFM, the global branch will pay more attention to anomalous mouth movements and the local branch will gain more information about the global context. Finally, a multi-task learning scheme is designed to optimize the network with both the global and local information. Extensive experiments have demonstrated that our approach achieves better detection results compared with several SOTA methods, especially in detecting video forgeries manipulated by unseen methods.
引用
收藏
页码:4027 / 4040
页数:14
相关论文
共 50 条
  • [31] Deepfake Video Detection Method Improved by GRU and Involution
    Liu, Yalin
    Lu, Tianliang
    Computer Engineering and Applications, 2023, 59 (22) : 276 - 283
  • [32] Deepfake Video Detection Based on MesoNet with Preprocessing Module
    Xia, Zhiming
    Qiao, Tong
    Xu, Ming
    Wu, Xiaoshuai
    Han, Li
    Chen, Yunzhi
    SYMMETRY-BASEL, 2022, 14 (05):
  • [33] An Investigation into the Utilisation of CNN with LSTM for Video Deepfake Detection
    Tipper, Sarah
    Atlam, Hany F.
    Lallie, Harjinder Singh
    APPLIED SCIENCES-BASEL, 2024, 14 (21):
  • [34] DeepMark: A Scalable and Robust Framework for DeepFake Video Detection
    Tang, Li
    Ye, Qingqing
    Hu, Haibo
    Xue, Qiao
    Xiao, Yaxin
    Li, Jin
    ACM TRANSACTIONS ON PRIVACY AND SECURITY, 2024, 27 (01)
  • [35] Hierarchical Contrastive Inconsistency Learning for Deepfake Video Detection
    Gu, Zhihao
    Yao, Taiping
    Chen, Yang
    Ding, Shouhong
    Ma, Lizhuang
    COMPUTER VISION, ECCV 2022, PT XII, 2022, 13672 : 596 - 613
  • [36] Advancements and Challenges in Deepfake Video Detection: A Comprehensive Review
    Gupta, Deepti
    Jain, Bhawna
    Sharma, Arun
    SMART TRENDS IN COMPUTING AND COMMUNICATIONS, VOL 1, SMARTCOM 2024, 2024, 945 : 353 - 369
  • [37] Deepfake video detection using deep learning algorithms
    Korkmaz, Sahin
    Alkan, Mustafa
    JOURNAL OF POLYTECHNIC-POLITEKNIK DERGISI, 2023, 26 (02): : 855 - 862
  • [38] DeepFake Video Detection: A Time-Distributed Approach
    Singh A.
    Saimbhi A.S.
    Singh N.
    Mittal M.
    SN Computer Science, 2020, 1 (4)
  • [39] Unsupervised Learning-Based Framework for Deepfake Video Detection
    Zhang, Li
    Qiao, Tong
    Xu, Ming
    Zheng, Ning
    Xie, Shichuang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 4785 - 4799
  • [40] Cascaded Network Based on EfficientNet and Transformer for Deepfake Video Detection
    Deng, Liwei
    Wang, Jiandong
    Liu, Zhen
    NEURAL PROCESSING LETTERS, 2023, 55 (06) : 7057 - 7076