Abstract
This paper presents a method of feature extraction for the automatic recognition of voiceless unaspirated stop consonants in Mandarin speech. The features are derived from the spectrographic acoustic patterns of syllable-initial voiceless unaspirated stops /p,t,k/, which include the burst spectrum, the formant transition, and the voice onset time. A normalization process for the second and the third formants at the voice onset is proposed. Based on these derived features, Bayes classifiers and a layered neural net are applied to classify the places of articulation of these stop consonants. The experiments show that the derived features are robust and efficient for speaker-independent speech recognition, and the neural net is a preferable choice in the classification of these stops in multiple contexts.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.