Abstract
Lip-syncing is a process of speech assimilation with the lip motions of a virtual character. A virtual talking character is a challenging task because it should provide control on all articulatory movements and must be synchronized with the speech signal. This study presents a virtual talking character system aimed to speeding and easing the visual talking process as compared to the previous techniques using the blend shapes approach. This system constructs the lip-syncing using a set of visemes for reduced phonemes set by a new method named Prophone. This Prophone depend on the probability of appearing the phoneme in the sentence of English Language. The contribution of this study is to develop real-time automatic talking system for English language based on the concatenation of the visemes, followed by presenting the results that was evaluated by the phoneme to viseme table using the Prophone.
Highlights
The most natural way to conveying the ideas of the personality of a virtual character is through the speech
All talkers have influenced the production of a given phoneme but not all visible, different phonemes may be associated with the same viseme
Visual talking system is a challenging system that creates 3D character based on text or audio or both input to obtain a virtual character like a human’s facial movements
Summary
The most natural way to conveying the ideas of the personality of a virtual character is through the speech. The challenging task in virtual talking animation is the selection of the visemes synchronized with the audio and the modelling of articulation. In a Talking character to synchronize the facial movements phonetic speech, combining phonemes, rather than the with an audio of utterance becomes impractical if the actual letters, creates words. Visemes is the visual Prophone is represented as a set of animation using a phoneme that describes the facial and the pose during canonical set of face poses. The proposed system is a clear distinction The evaluation of the proposed system was carried out between the talk processes of the input utterance and using three sentences presented: the animation of the virtual character. Our system used Automatic Speech Recognition (ASR) and text to speech synthesis (TTS)
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: Research Journal of Applied Sciences, Engineering and Technology
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.