Abstract

Fundamental frequency estimation is one of the most important issues in the field of speech processing. An accurate estimate of the fundamental frequency plays a key role in the field of speech and music analysis. So far, various methods have been proposed in the time- and frequency-domain. However, the main challenge is the strong noises in speech signals. In this paper, to improve the accuracy of fundamental frequency estimation, we propose a method for optimal combination of fundamental frequency estimation methods, in noisy signals. In this method, to discriminate voiced frames from unvoiced frames in a better way, the Voiced/Unvoiced (V/U) scores of four pitch detection methods are combined linearly. These methods are: Autocorrelation, Yin, YAAPT and SWIPE. After identifying the Voiced/Unvoiced label of each frame, the fundamental frequency (F 0 ) of the frame is estimated using the SWIPE method. The optimal coefficients for linear combination are determined using the regularized least squares method with Tikhonov regularization. To evaluate the proposed method, 10 speech files (5 female and 5 male voices) are selected from the PTDB-TUG standard database and the results are presented in terms of SDFPE, GPE, VDE, PTE and FFE standard error criteria. The results indicate that our proposed method relatively reduced the aforementioned criteria (averaged in various SNRs) by 27.13%, 22.14%, 17.40%, and 26.74% respectively, which demonstrate the effectiveness of the proposed method in comparison to state-of-the-art methods.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.