Abstract

Current era is to make the interaction between humans and their artificial partners (Computers) and make communication easier and more reliable. One of the actual tasks is the use of vocal interaction. Speech recognition may be improved by visual information of human face. In literature, the lip shape and its movement are referred to as lip reading. Lip reading computing plays a vital role in automatic speech recognition and is an important step towards accurate and robust speech recognition. In this paper we consider our approach towards automatic lip reading in detail by using Active Appearance Model (AAM) and Hidden Markov Model (HMM). Appearance-based methods consider the raw image for feature extraction. AAM uses for detection of speaker's face features and feature points from faces automatically by appearance parameters of a speaker's lips. Hidden Markov Model (HMM) approach is used for Lip recognition. In this paper, we will describe the visual features based on shape and appearance descriptions and detailed information about the automatic lip reading system. Individual characteristics and efficiency of both approaches towards the Lip reading is also discuss. The efficiency of both approaches (AAM and HMM) will be evaluated individually by characteristics and their performances are then compared and discuss. In this paper we will show the recognition process of common visual features (i.e. selection of lip and extraction of movement) for an improved lip reading.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call