Convolutional long short-term memory-based approach for deepfakes detection from videos

被引:0
作者
Marriam Nawaz
Ali Javed
Aun Irtaza
机构
[1] UET Taxila,Department of Software Engineering
[2] UET Taxila,Department of Computer Science
来源
Multimedia Tools and Applications | 2024年 / 83卷
关键词
CNN; Deepfakes; Bi-LSTM; Deep learning; Multimedia forensic;
D O I
暂无
中图分类号
学科分类号
摘要
The great development in the area of Artificial Intelligence (AI) has introduced tremendous advancements in information technology. Moreover, the introduction of lightweight machine learning (ML) techniques allows the applications to work with limited storage and processing power. Deepfakes is among the most famous type of such applications of this era which generates a large amount of fake and modified audiovisual data. The creation of such fake data has introduced a serious risk to the security and confidentiality of humans all around the globe. Accurate detection and classification of actual and deepfakes content is a challenging task due to the progression of Generative adversarial networks (GANs) which produce such convincing manipulated content that it’s impossible for people to recognize it through their naked eyes. In this work, we have presented deep learning (DL)-based approach namely the convolutional long short-term memory (C-LSTM) method for deepfakes detection from videos. More specifically, the spatial information from the input sample is calculated by employing various pre-trained models like VGG16, VGG19, ResNet50, XceptionNet, and GoogleNet, DenseNet. Further, we have proposed a novel feature descriptor called the Dense-Swish-Net121. Whereas the Bi-LSTM model is utilized to compute the temporal information. Lastly, the results are predicted based on both the frame level and temporal level information to make the final decision. A detailed comparison of all CNN models with the Bi-LSTM approach is performed and has confirmed through the reported results that the proposed Dense-Swish-Net121 with Bi-LSTM approach performs well for deepfakes detection.
引用
收藏
页码:16977 / 17000
页数:23
相关论文
共 65 条
[1]  
Albahli S(2021)Recognition and Detection of Diabetic Retinopathy Using Densenet-65 Based Faster-RCNN Comput Mater Contin 67 1333-1351
[2]  
Nazir T(2017)Video saliency detection via spatial-temporal fusion and low-rank coherency diffusion IEEE Trans Image Process 26 3156-3170
[3]  
Irtaza A(2021)Exploring rich and efficient spatial temporal interactions for real-time video salient object detection IEEE Trans Image Process 30 3995-4007
[4]  
Javed A(2019)Improved robust video saliency detection based on long-term spatial-temporal information IEEE Trans Image Process 29 1090-1100
[5]  
Chen C(2023)AVFakeNet: A unified end-to-end Dense Swin Transformer deep learning model for audio–visual​ deepfakes detection Appl Soft Comput 136 110124-18478
[6]  
Li S(2023)DFGNN: An interpretable and generalized graph neural network for deepfakes detection Expert Syst Appl 222 119843-1167
[7]  
Wang Y(2021)Detecting DeepFake, FaceSwap and Face2Face facial forgeries using frequency CNN Multimed Tools Appl 80 18461-2327
[8]  
Qin H(2022)Exposing deepfakes using a deep multilayer perceptron–convolutional neural network model Int J Inf Manage Data Insights 2 100054-4026
[9]  
Hao A(2019)Accurate and robust video saliency detection via self-paced diffusion IEEE Trans Multimedia 22 1153-10371
[10]  
Chen C(2020)A plug-and-play scheme to adapt image saliency deep model for video data IEEE Trans Circuits Syst Vid Technol 31 2315-1944