Abstract

A new algorithm is developed for voiced-unvoiced speech discrimination in noise. Short segments of speech are modeled as a sum of basis functions from a Gabor dictionary. In each iteration, a Gabor atom is fitted (using the matching pursuit algorithm) to the residual obtained by subtracting the best-fit Gabor atom from the previous residual. Multiple discriminant analysis is used to reduce the dimensionality of the vector of Gabor coefficients to give a low-dimensional feature vector for classification. A radial basis function neural network is trained on the reduced feature vector set to discriminate between voiced and unvoiced speech/silence segments. On a database of 62 sentences in 5-dB SNR speech-shaped noise, 84% correct classification accuracy was obtained.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.