Abstract

MFCC (Mel-Frequency Cepstral Coefficients) is a kind of traditional speech feature widely used in speech recognition. The error rate of the speech recognition algorithm using MFCC and CDHMM is known to be very low in a clean speech environment, but it increases greatly in a noisy environment, especially in the white noisy environment. We propose a new kind of speech feature called the auditory spectrum based feature (ASBF) that is based on the second-order difference cochlear model of the human auditory system. This new speech feature can track the speech formants and the selection scheme of this feature is based on both the second-order difference cochlear model and primary auditory nerve processing model of the human auditory system. In our experiment, the performance of MFCC and ASBF are compared in both clean and noisy environments when left-to-right CDHMM with 6 states and 5 Gaussian mixtures is used. The experimental result shows that the ASBF is much more robust to noise than MFCC. When only 5 frequency components are used in ASBF, the error rate is approximately 38% lower than the traditional MFCC with 39 parameters in the condition of S/N=10 dB with white noise.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call