Abstract
Speech emotion recognition(SER) task based on acoustic wave always has a large number of features, which brings great difficulties to improve the accuracy of recognition. In this paper, we propose a new speech emotion recognition method, which is based on depth wavefield extrapolation and improved wave physics model (DWE-WPM). The method can improve loss accuracy and feature explosion problem when extracting the features. The schema comes from the wave physics system. After extrapolating the wavefield with a fixed-step depth, we inject the reconstructed waveform into DWE-WPM to simulate the information mining process of Long Short-Term Memory Recurrent Neural Network(LSTM), and then fuse the output features of this model with the sorted HSF features. Finally, the integrated features are injected into BiMLSTM to automatically complete the SER task. Massive experiments were carried out on the emotion corpus of interactive emotional dyadic motion capture (IEMOCAP). The experimental results showed that the weighted average (UA) accuracy of the proposed method can be improved by 21%, which was better than the existing methods of SER from raw wave. The method proposed in the paper proved the effective for SER task.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.