Abstract

Abstract In this paper, the extracted acoustic features are processed using the MFCC method to classify the sound recognition. The linear spectrum of the sound is mapped onto the Mel FPP coefficients and Mel nonlinear spectrum, respectively, and the MFCC is obtained by cepstrum processing. The Mel FPP coefficients and Gaussian mixture model are combined to create the acoustic model. On this basis, the art and emotional skills of vocal singing are explored, and simulation and empirical experiments are set up to analyze the effect of the model constructed in this paper. The experimental results show that the model constructed in this paper tends to stabilize when the number of iterations is 2, the final accuracy rate stabilizes near 0.9, and the model is effective. Different vocal types are recognized by the model constructed in this paper, and the recognition rate of emotional expression of all vocal types is above 0.7, and the model recognition rate is high. The accuracy and recall classification for emotional expression techniques are 0.73 and 0.79, resulting in a more balanced evaluation overall.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call