Sign language serves as a vital communication tool for individuals with hearing and speech impairments. However, a significant barrier exists when others do not possess a strong understanding of sign language often requiring interpreters to facilitate communication. To reduce the reliance on human interpreters, this research aims to develop an intelligent system capable of recognizing and translating sign language gestures into meaningful grammatically correct sentences. The proposed system processes both images and videos to interpret gestures, converting them into text that can be translated into any desired language. The system leverages state-of-the-art deep learning techniques, such as Convolutional Neural Networks (CNNs) for identifying key features in images and Recurrent Neural Networks (RNNs) for understanding temporal sequences in video data. By employing these advanced neural networks, the model is able to comprehend hand movements, facial expressions and other non-verbal cues to construct coherent sentences. Additionally, the system integrates natural language processing (NLP) to refine the output, ensuring the resulting sentences are grammatically correct. Our approach addresses common challenges such as differentiating between subtle hand gestures and reducing the impact of environmental noise in images. This solution holds the potential to significantly enhance communication for the hearing and speech impairedffering an efficient interpreter-free method of translating sign language into widely spoken languages