Abstract

Audio-visual emotion recognition is the research of identifying human emotional states by combining the audio modality and the visual modality simultaneously, which plays an important role in intelligent human-machine interactions. With the help of deep learning, previous works have made great progress for audio-visual emotion recognition. However, these deep learning methods often require a large amount of data for training. In reality, data acquisition is difficult and expensive, especially for the multimodal data with different modalities. As a result, the training data may be in the low-data regime, which cannot be effectively used for deep learning. In addition, class imbalance may occur in the emotional data, which can further degrade the performance of audio-visual emotion recognition. To address these problems, we propose an efficient data augmentation framework by designing a multimodal conditional generative adversarial network (GAN) for audio-visual emotion recognition. Specifically, we design generators and discriminators for audio and visual modalities. The category information is used as their shared input to make sure our GAN can generate fake data of different categories. In addition, the high dependence between the audio modality and the visual modality in the generated multimodal data is modeled based on Hirschfeld-Gebelein-Rényi (HGR) maximal correlation. In this way, we relate different modalities in the generated data to approximate the real data. Then, the generated data are used to augment our data manifold. We further apply our approach to deal with the problem of class imbalance. To the best of our knowledge, this is the first work to propose a data augmentation strategy with a multimodal conditional GAN for audio-visual emotion recognition. We conduct a series of experiments on three public multimodal datasets, including eNTERFACE’05, RAVDESS, and CMEW. The results indicate that our multimodal conditional GAN has high effectiveness for data augmentation of audio-visual emotion recognition.

Highlights

  • Tsinghua Shenzhen International Graduate School, Tsinghua University, Shenzhen 518055, China; Abstract: Audio–visual emotion recognition is the research of identifying human emotional states by combining the audio modality and the visual modality simultaneously, which plays an important role in intelligent human–machine interactions

  • Our proposed multimodal conditional generative adversarial network (GAN) is a generalization of existing GANs for data augmentation to improve the performance of audio–visual emotion recognition

  • We find the following summarizations: (1) Our approach achieves the highest performance compared to other methods, which shows that the data generated using our multimodal conditional GAN can significantly benefit audio–visual emotion recognition

Read more

Summary

Introduction with regard to jurisdictional claims in

The task of emotion recognition is to detect human affective states. It is crucial for affect-related human–machine interactions, which has attracted a lot of attention from researchers [1,2,3,4,5,6,7,8]. Our proposed multimodal conditional GAN is a generalization of existing GANs for data augmentation to improve the performance of audio–visual emotion recognition. Additional category information is used as their shared input to generate fake data of different categories It is shown in [17,47,48,49] that in the real multimodal data, the audio modality and the visual modality are highly dependent, which is beneficial to emotion recognition. We conduct experiments on three public multimodal datasets to show that our multimodal conditional GAN can be effectively used for data augmentation of audio–visual emotion recognition. To the best of our knowledge, this is the first work to propose an efficient data augmentation approach with a multimodal conditional GAN for audio–visual emotion recognition.

Multimodal Learning
Overview
Proposed Multimodal Conditional GAN
DNN Classifier
Datasets
Networks
Implementation Details
Experiment Results
Method
Conclusions
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call