Abstract

The rapid increase in the importance of human-machine interaction and the accelerating pace of life pose various challenges for the creators of digital environments. Continuous improvement of human-machine interaction requires precise modeling of the physical and emotional state of people. By implementing emotional intelligence in machines, robots are expected not only to recognize and track emotions when interacting with humans, but also to respond and behave appropriately. The machine should match its reaction to the mood of the user as precisely as possible. Music generation with a given emotion can be a good start to fulfilling such a requirement. This article presents the process of building a system generating music content of a specified emotion. As the emotion labels, four basic emotions: happy, angry, sad, relaxed, corresponding to the four quarters of Russell’s model, were used. Conditional variational autoencoder using a recurrent neural network for sequence processing was used as a generative model. The obtained results in the form of the generated music examples with a specific emotion are convincing in their structure and sound. The generated examples were evaluated with two methods, in the first using metrics for comparison with the training set and in the second using expert annotation.

Highlights

  • More and more devices and machines enter our everyday life

  • The implemented emotional intelligence machines and robots are expected to recognize and track emotions when interacting with humans and to respond and behave appropriately to the actual human

  • A much smaller number of notes is in Figs. 10c and 10d, examples with emotions sad (e3) and relaxed (e4)

Read more

Summary

Introduction

Human-machine interaction can be encountered in industry. It started more than half a century ago with industrial robots [1]. They were joined by increasingly complex and multifunctional information and vending machines, and today this interaction is almost everywhere, e.g. a great number of people are increasingly using e-assistants like Amazon Alexa and Google Assistant. The importance of human-machine interaction on the one hand and customer expectations on the other set quality requirements for new machine generation. The implemented emotional intelligence machines and robots are expected to recognize and track emotions when interacting with humans and to respond and behave appropriately to the actual human

Objectives
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call