Abstract

The automatic analysis of speech to detect affective states may improve the way users interact with electronic devices. However, the analysis only at the acoustic level could be not enough to determine the emotion of a user in a realistic scenario. In this paper we analyzed the spontaneous speech recordings of the FAU Aibo Corpus at the acoustic and linguistic levels to extract two sets of features. The acoustic set was reduced by a greedy procedure selecting the most relevant features to optimize the learning stage. We compared two versions of this greedy selection algorithm by performing the search of the relevant features forwards and backwards. We experimented with three classification approaches: Naive-Bayes, a support vector machine and a logistic model tree, and two fusion schemes: decision-level fusion, merging the hard-decisions of the acoustic and linguistic classifiers by means of a decision tree; and featurelevel fusion, concatenating both sets of features before the learning stage. Despite the low performance achieved by the linguistic data, a dramatic improvement was achieved after its combination with the acoustic information, improving the results achieved by this second modality on its own. The results achieved by the classifiers using the parameters merged at feature level outperformed the classification results of the decision-level fusion scheme, despite the simplicity of the scheme. Moreover, the extremely reduced set of acoustic features obtained by the greedy forward search selection algorithm improved the results provided by the full set.

Highlights

  • ONE of the goals of human-computer interaction (HCI) is the improvement of the user experience, trying to make this interaction closer to human-human communication

  • Both modalities were combined at the decision level and at the feature level to compare the performance of different classification approaches using both procedures

  • The results obtained by the dataset created by the FW search procedure are shown at the top and the results achieved by the BW search dataset are shown at the bottom

Read more

Summary

INTRODUCTION

ONE of the goals of human-computer interaction (HCI) is the improvement of the user experience, trying to make this interaction closer to human-human communication. Because in a realistic scenario the analysis of acoustic information could be not enough to carry out the task of emotion recognition from speech [9] the linguistic modality could improve an only-acoustic study. In this article, both modalities were combined at the decision level and at the feature level to compare the performance of different classification approaches using both procedures.

Corpus Description
Acoustic Parameterization
Linguistic Parameterization
Methodology
Feature Selection Process
Dataset Pre-processing
Decision-Level Fusion
Feature-Level Fusion
RESULTS
CONCLUSION

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.