SARB-DF: A Continual Learning Aided Framework for Deepfake Video Detection Using Self-Attention Residual Block

被引:0
作者
Prathibha, P. G. [1 ]
Tamizharasan, P. S. [1 ]
Panthakkan, Alavikunhu [2 ]
Mansoor, Wathiq [2 ]
Al Ahmad, Hussain [2 ]
机构
[1] Birla Inst Technol & Sci Pilani, Dubai Campus, Dubai, U Arab Emirates
[2] Univ Dubai, Dept Engn & IT, Dubai, U Arab Emirates
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Deepfakes; Feature extraction; Convolutional neural networks; Computational modeling; Continuing education; Forgery; Faces; Data models; Adaptation models; Transformers; Deepfake; continual learning; self-attention; uncertainty sampling;
D O I
10.1109/ACCESS.2024.3517170
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The creation and dissemination of deepfake videos have become increasingly prevalent nowadays, facilitated by advanced technological tools. These synthetic videos pose significant security challenges as they can spread misinformation and manipulation thereby undermining the digital media. Owing to the continuous generation of novel synthetic data, deepfake detection models must be regularly updated to enhance their generalization capabilities. In this research article, we propose a deepfake video detection system with self-attention mechanism and continual learning. A self-attention residual module is specifically introduced to extract detailed facial features. We enable the deepfake detection process with continual learning to improve detection capability and improve generalization. The framework uses weight regularization and a dynamic sample set to continuously learn and adapt to new synthetic data. We demonstrated our proposed approach on Xception-Net backbone with benchmark datasets such as Celeb-DF and Face-Forensics++ datasets. Experimental results shows AUC values of 99.26% on Celeb-DF and 99.67%, 93.57%, 99.78% and 90.00% on different categories on FaceForensics++ such as Deepfakes, Face2Face, FaceSwap and Neural Textures respectively.
引用
收藏
页码:189088 / 189101
页数:14
相关论文
共 72 条
  • [1] Verdoliva L., Media forensics and DeepFakes: An overview, IEEE J. Sel. Topics Signal Process., 14, 5, pp. 910-932, (2020)
  • [2] Karras T., Aittala M., Hellsten J., Laine S., Lehtinen J., Aila T., Training generative adversarial networks with limited data, Proc. Adv. Neural Inf. Process. Syst., 33, pp. 12104-12114, (2020)
  • [3] Bonettini N., Cannas E.D., Mandelli S., Bondi L., Bestagini P., Tubaro S., Video face manipulation detection through ensemble of CNNs, Proc. 25th Int. Conf. Pattern Recognit. (ICPR), pp. 5012-5019, (2021)
  • [4] Ahmed S.R., Sonuc E., Ahmed M.R., Duru A.D., Analysis survey on deepfake detection and recognition with convolutional neural networks, Proc. Int. Congr. Hum.-Comput. Interact., Optim. Robotic Appl. (HORA), pp. 1-7, (2022)
  • [5] Caldelli R., Galteri L., Amerini I., Del Bimbo A., Optical flow based CNN for detection of unlearnt deepfake manipulations, Pattern Recognit. Lett., 146, pp. 31-37, (2021)
  • [6] Mazaheri G., Roy-Chowdhury A.K., Detection and localization of facial expression manipulations, Proc. IEEE/CVF Winter Conf. Appl. Comput. Vis. (WACV), pp. 1035-1045, (2022)
  • [7] Park J., Kim Y.-G., Styleformer: Transformer based generative adversarial networks with style vector, Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 8983-8992, (2022)
  • [8] Xu Z., Zhou H., Hong Z., Liu Z., Liu J., Guo Z., Han J., Liu J., Ding E., Wang J., StyleSwap: Style-based generator empowers robust face swapping, Proc. Eur. Conf. Comput. Vis., pp. 661-677, (2022)
  • [9] Singh S., Sharma R., Smeaton A.F., Using GANs to synthesise minimum training data for deepfake generation
  • [10] Gerstner C.R., Farid H., Detecting real-time deep-fake videos using active illumination, Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. Workshops (CVPRW), pp. 53-60, (2022)