The rapid advancement of deep learning techniques, creating realistic multimedia content has become increasingly accessible, leading to the proliferation of DeepFake technology. DeepFake utilizes generative deep learning algorithms to produce or modify face features in a highly realistic manner, often making it challenging to differentiate between real and manipulated media. This technology, while beneficial in fields such as entertainment and education, also poses significant threats, including misinformation and identity theft. Consequently, detecting DeepFakes has become a critical area of research. In this paper, we propose a novel approach to DeepFake face detection by integrating Convolutional Neural Networks (CNN) with Long Short-Term Memory (LSTM) networks. Our method leverages the strengths of CNNs in spatial feature extraction and LSTMs in temporal sequence modeling to enhance detection accuracy. The CNN component captures intricate facial features, while the LSTM analyzes the temporal dynamics of video frames. We evaluate our model on several benchmark datasets, including Celeb-DF (v2), DeepFake Detection Challenge Preview, and FaceForensics++. Experimental results demonstrate that our hybrid CNN-LSTM model achieves state-of-the-art performance, surpassing existing methods in both accuracy and robustness. This study highlights the potential of combining CNN and LSTM architectures for effective DeepFake detection, contributing to the ongoing efforts to safeguard against digital media manipulation