Abstract

In order to recognise speech in a background of other sounds, human listeners must solve two perceptual problems. First, the mixture of sounds reaching the ears must be parsed to recover a description of each acoustic source, a process termed 'auditory scene analysis'. Second, recognition of speech must be robust even when the acoustic evidence is missing due to masking by other sounds. This paper describes an automatic speech recognition system that addresses both of these issues, by combining a neural oscillator model of auditory scene analysis with a framework for 'missing data' recognition of speech.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.