Abstract
A system of communication called sign language makes use of visual motions and signals. The only form of communication for the deaf and dumb community and others with hearing impairments is sign language. Understanding sign language is so much difficult for a normal person. As a result, connecting with the wider public has always been extremely difficult for the minority community. In this study, we suggested a novel deep learning-based method for identifying sign language that can help normal and deaf individuals communicate more easily. In order to identify real-time sign language, we first created a dataset with 11 sign terms. Our bespoke CNN model was trained using these sign words. Prior to the CNN model being trained, we preprocessed the dataset. Our results show that the customized CNN model can attain the greatest accuracy of 98.6%.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: Shanlax International Journal of Arts, Science and Humanities
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.