Abstract

Creating a computational device to identify human emotions via voice analysis represents a notable achievement in the sector of human-computer interaction, especially within the healthcare domain. We propose a new light-weight model for addressing challenges of emotions recognition. The model works based on CNN with change of kernel processing. The proposed model performs a direct matching to recognize speech emotions of different eight categories using a statistical model named Analysis of Variance (ANOVA) as kernel for features extraction and Cosine Similarity Measurement (CSM) as activation function for CNN model. This proposed model contains eight-folded single-layered intermediate neurons, and each neuron can segregate speech emotion pattern using CSM from the voice convergence matrix to explore a part of the solution from the whole solution. Experiment results demonstrates that the proposed model outperforms compared with multiple layered existing CNN methods in identifying the emotional state of a speaker.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call