Abstract

Our project falls under the domain of Natural Language Processing and Machine Learning. With the advent of digitization, it became necessary to preserve data produced by humans. Voice and text are how we communicate. Visual impairment can be defined as partial or complete loss of vision. The primary goal of this project is to extract text from braille text images, so as to store it in a drive for later use. Although there are various ways to implement this, we'll use Deep Learning. Visually impaired people face numerous difficulties like interacting with the surroundings, navigation, etc. Different machine learning models like CNN, etc. have been used to solve this problem but usually have a high cost of detection in terms of computational time and hence they are not suitable for real-time applications [1]. To meet the project objective, we studied about Convolutional Neural Networks. The plan for picking this field is that it empowers machines to see the world as people do, see it likewise and even utilize the information for a huge number of undertakings like Image Recognition, Image Analysis and Classification, Media Recreation, Recommendation Systems, and so on [2]. This paper proposes a customized CNN model based on Braille Character dataset which has 26 different classes of all alphabets. This model uses a feature extractor that recognizes relevant features corresponding to each image, and then uses a classifier network to classify images into a set of classes. This model will take an image input, then pass those individual images to our classifier and then detect the braille language characters present in that image and give a translated English text and audio output. The final accuracy we obtained was 0.9615. These results show the model's robustness against factors like low light illumination.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call