Abstract

Sign language is the most natural and effective way for communications among deaf and normal people. American Sign Language (ASL) alphabet recognition (i.e. fingerspelling) using marker-less vision sensor is a challenging task due to the difficulties in hand segmentation and appearance variations among signers. Existing color-based sign language recognition systems suffer from many challenges such as complex background, hand segmentation, large inter-class and intra-class variations. In this paper, we propose a new user independent recognition system for American sign language alphabet using depth images captured from the low-cost Microsoft Kinect depth sensor. Exploiting depth information instead of color images overcomes many problems due to their robustness against illumination and background variations. Hand region can be segmented by applying a simple preprocessing algorithm over depth image. Feature learning using convolutional neural network architectures is applied instead of the classical hand-crafted feature extraction methods. Local features extracted from the segmented hand are effectively learned using a simple unsupervised Principal Component Analysis Network (PCANet) deep learning architecture. Two strategies of learning the PCANet model are proposed, namely to train a single PCANet model from samples of all users and to train a separate PCANet model for each user, respectively. The extracted features are then recognized using linear Support Vector Machine (SVM) classifier. The performance of the proposed method is evaluated using public dataset of real depth images captured from various users. Experimental results show that the performance of the proposed method outperforms state-of-the-art recognition accuracy using leave-one-out evaluation strategy.

Highlights

  • Sign and gestures are considered as the most natural way to convey messages among people through body movements [1]–[4]

  • EXPERIMENTAL RESULTS we examine the performance of the proposed system for signer independent recognition scenario

  • This paper proposes a new efficient method for user independent American fingerspelling recognition based on depth images and Principal Component Analysis Network (PCANet) features

Read more

Summary

Introduction

Sign and gestures are considered as the most natural way to convey messages among people through body movements [1]–[4]. Though signs and gestures are classified as a non-verbal communication, they can effectively deliver the communicating messages among deaf and hearing-impaired people [5]. The most widely used method of conveying words/vocabularies using body gestures is sign language. A plenty of research works in automatic Sign Language Recognition (SLR) have been started two decades ago. The associate editor coordinating the review of this article and approving it for publication was Farid Boussaid. Many techniques based on different sensor types have been developed. These approaches employed variety of methods based on the combination of multiple sensors, machine learning, pattern recognition and image analysis techniques

Methods
Results
Conclusion

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.