Abstract

In the human language communication process, facial expression and lip shape change of the teller contains extremely rich language information. Previous research about lip-reading recognition focused on proposing the theoretical method to analyze rules of lip shapes. However, little evidence appears that the real lip reading recognition system was implemented. For getting moving lip features, principal component analysis and mouth changing rate were put forward by this paper. We implement a lip language recognition system by using image processing technology, neural network algorithm, and database to help people and computer understand lip language. The dynamic image variation of the lip shape changing could be detected in this system. The results indicated that this method could recognize effectively and correctly 62 Chinese words with 55 users and practically enhanced the lip language recognition.

Highlights

  • Owing to the fast growing technology of image processing and computer hardware, computer input interfaces have been becoming more and more important and necessary within the past few decades

  • Most of these articles only discuss the methods of lip shape feature grabbing or tracking, not a complete lip reading recognition system (Chiang et al, 2003; Yao et al.,2010)

  • The Principal Component Analysis (PCA) method could decrease the dimensions of raw data allowing the image data to be condensed without losing its major features

Read more

Summary

Introduction

Owing to the fast growing technology of image processing and computer hardware, computer input interfaces have been becoming more and more important and necessary within the past few decades. Supposing a user sits in front of the CCD camera, this system could track the lip position from the user’s face to recognize whether the lip changing of a word pronunciation is correct or not. This system could be used as a lip shape control interface. According to the research results of the University of Manchester (Bauman, 2003), the hearing impaired subjects could only recognize 21% of spoken language If they used a hearing aid, the recognition rate of speaking could be increased to 65%. The dynamic image variation of the lip shape changing could be detected in this system for using as a lip shape control interface

Image Processing Steps
Position Detection and Lip Tracking
RGB and YCbCr Color space transformation
The segmentation of skin portion
The enhancement of lip feature
Region of Interest
Neural Network
Experimental Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.