Abstract

Acoustic cues are robust elements that can be used to infer information contained in the speech signal, such as underlying linguistic distinctive features and the words intended by the speaker (Stevens JASA 2002). Yet, most current automatic speech recognition systems do not take advantage of a feature-cue-based framework for signal analysis. In this project, a set of common acoustic cues has been explicitly modeled by Gaussian mixture models. This set of acoustic cues can provide evidence for the overall phoneme and word sequences of an utterance. The extracted cues and their values can also determine a speaker’s linguistic production pattern, i.e., the systematic context-governed modifications in surface-phonetic form that occur pervasively in conversational speech. The simple Gaussian mixture model representation structure reduces the need for extensive amounts of training data, in contrast to conventional schemes based on large neural networks.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call