Abstract

This study proposes a system for recognizing static gestures in Taiwan sign languages (TSL), using 3D data and neural networks trained to completion. The 3D hand gesture data is acquired from VICON, and are processed and converted to features that are fed to a neural network. The extracted features are invariant for occlusion, rotation, scaling and translation of hand. Experimental results indicate that the proposed system can recognize the 20 static hand gestures in Taiwan sign language with an average accuracy of 96.58%. Additionally, the difference in recognition rate between training set and test set is 3.98%, showing that the proposed system is robust.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call