Abstract

Artificial intelligence advancements have led to the development of deepfake technology, which seriously jeopardises the integrity of visual media material. Robust detection algorithms are becoming more and more necessary as deepfake creation techniques become more complex. This study combines Long Short-Term Memory (LSTM) networks with Convolutional Neural Networks (CNNs) to present a novel method for deepfake identification. The suggested CNN+LSTM architecture makes use of LSTMs' temporal modelling capabilities and CNNs' spatial feature extraction capabilities. While the LSTM component analyses the temporal connections between frames to identify patterns suggestive of deepfake manipulation, the CNN component concentrates on capturing local features and patterns in individual frames. The combination of these two networks improves the model's capacity to identify minute anomalies and inconsistencies that are indicative of deepfake content. To extract frame-level characteristics, we use Res-Next Convolutional Neural Networks. These attributes are then used to train a Recurrent Neural Network (RNN) based on Long Short-Term Memory (LSTM) to determine whether a video has been manipulated, i.e., whether it is a deepfake or a genuine video. We intend to train our deepfake detection model on a varied set of public datasets in order to improve its real-time performance. We improve the model's adaptability by learning features from different photos. Face-Forensic++, Deepfake Detection Challenge, and Celeb-DF datasets are used to extract videos. Furthermore, to assure competitive performance in real-world scenarios, our model will be assessed against a large amount of real-time data, including the YouTube dataset. Key Words: Temporal modelling, Deepfake technology , Long Short-Term Memory (LSTM) networks, Convolutional Neural Networks (CNNs)

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call