Abstract

Deafness does not restrict its negative effect on the person’s hearing, but rather on all aspect of their daily life. Moreover, hearing people aggravated the issue through their reluctance to learn sign language. This resulted in a constant need for human translators to assist deaf person which represents a real obstacle for their social life. Therefore, automatic sign language translation emerged as an urgent need for the community. The availability and the widespread use of mobile phones equipped with digital cameras promoted the design of image-based Arabic Sign Language (ArSL) recognition systems. In this work, we introduce a new ArSL recognition system that is able to localize and recognize the alphabet of the Arabic sign language using a Faster Region-based Convolutional Neural Network (R-CNN). Specifically, faster R-CNN is designed to extract and map the image features, and learn the position of the hand in a given image. Additionally, the proposed approach alleviates both challenges; the choice of the relevant features used to encode the sign visual descriptors, and the segmentation task intended to determine the hand region. For the implementation and the assessment of the proposed Faster R-CNN based sign recognition system, we exploited VGG-16 and ResNet-18 models, and we collected a real ArSL image dataset. The proposed approach yielded 93% accuracy and confirmed the robustness of the proposed model against drastic background variations in the captured scenes.

Highlights

  • Gesturing is one of the earliest forms of human communication

  • The facial expression and the body movement compose the non-manual components. Such sign language is perceived as a non-verbal communication way that is mainly intended to ease the communication for the Deaf and Hard of Hearing (DHH) persons

  • In order choose the relevant visual descriptors and enhance the segmentation accuracy, we propose to design and implement a novel Arabic Sign Language recognition system based on the Faster Region Convolutional neural network (R-Convolutional Neural Network (CNN))

Read more

Summary

Introduction

Deaf and Hard of Hearing (DHH) people are the predominant users of the officially recognized sign language which consists of alphabets, numbers, and words typically used to communicate within and outside their community. A sign language consists of; (i) manual components, and (ii) non-manual component. The configuration, the position, and the movement of the hands form the manual components. The facial expression and the body movement compose the non-manual components. Such sign language is perceived as a non-verbal communication way that is mainly intended to ease the communication for the DHH persons. Approximately 466 million people who suffer from a moderate to profound hearing loss struggle with communication daily. Deaf people cannot be considered as a linguistic minority which the language can be neglected

Objectives
Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call