Abstract

Speech emotion recognition(SER) task based on acoustic wave always has a large number of features, which brings great difficulties to improve the accuracy of recognition. In this paper, we propose a new speech emotion recognition method, which is based on depth wavefield extrapolation and improved wave physics model (DWE-WPM). The method can improve loss accuracy and feature explosion problem when extracting the features. The schema comes from the wave physics system. After extrapolating the wavefield with a fixed-step depth, we inject the reconstructed waveform into DWE-WPM to simulate the information mining process of Long Short-Term Memory Recurrent Neural Network(LSTM), and then fuse the output features of this model with the sorted HSF features. Finally, the integrated features are injected into BiMLSTM to automatically complete the SER task. Massive experiments were carried out on the emotion corpus of interactive emotional dyadic motion capture (IEMOCAP). The experimental results showed that the weighted average (UA) accuracy of the proposed method can be improved by 21%, which was better than the existing methods of SER from raw wave. The method proposed in the paper proved the effective for SER task.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call