Abstract

One of the best frequency trackers to date is the human (mammalian) auditory system, which has evolved through millions of years of resolving classification problems. It is a versatile, elegant and powerful sound processing unit. It excels in detecting, estimating, and classifying multiple targets simultaneously even in noisy environments. Hence, mimicking even some features of the auditory system could be beneficial in developing superior frequency tracking and classification algorithms. An auditory inspired adaptive synchrony capture filterbank (SCFB) signal processing architecture for tracking signal frequency components was proposed in a related paper [JASA (2013)]. The SCFB architecture consists of a fixed array of traditional, passive linear, gammatone filters in cascade with a bank of three adaptively tunable bandpass filters that form a frequency-discriminator-loop (FDL). The SCFB exhibits many desirable properties for processing speech, music, and other complex sounds. In recent work (Dec 2021), the algorithm was modified using adaptive tuning parameters, and a generalized way to determine/suppress voiced and unvoiced (silent) regions. This modified algorithm estimates frequencies with higher accuracy even in the presence of closely spaced input tones. Preliminary analysis with synthetic, human speech and humpback/whale-call signals demonstrates that the revised algorithm performs well. This talk will focus on the latest updates.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call