Abstract

In traditional music composition, the composer has a special knowledge of music and combines emotion and creative experience to create music. As computer technology has evolved, various music-related technologies have been developed. To create new music, a considerable amount of time is required. Therefore, a system is required that can automatically compose music from input music. This study proposes a novel melody composition method that enhanced the original generative adversarial network (GAN) model based on individual bars. Two discriminators were used to form the enhanced GAN model: one was a long short-term memory (LSTM) model that was used to ensure correlation between the bars, and the other was a convolutional neural network (CNN) model that was used to ensure rationality of the bar structure. Experiments were conducted using bar encoding and the enhanced GAN model to compose a new melody and evaluate the quality of the composition melody. In the evaluation method, the TFIDF algorithm was also used to calculate the structural differences between four types of musical instrument digital interface (MIDI) file (i.e., randomly composed melody, melody composed by the original GAN, melody composed by the proposed method, and the real melody). Using the TFIDF algorithm, the structures of the melody composed were compared by the proposed method with the real melody and the structure of the traditional melody was compared with the structure of the real melody. The experimental results showed that the melody composed by the proposed method had more similarity with real melody structure with a difference of only 8% than that of the traditional melody structure.

Highlights

  • The composer has a special knowledge of music and combines emotion and creative experience to create music

  • Music can be divided into several parts that are recombined based on special knowledge of musical composition to create new music

  • Continuous recurrent neural networks based on the generative adversarial network (GAN) (C-RNN-GAN) compose music by building a GAN model with two long short-term memory (LSTM), composing the melodies [11]

Read more

Summary

Introduction

The composer has a special knowledge of music and combines emotion and creative experience to create music. As computer technology has evolved, various music-related technologies have been developed, but most of them have been focused on music editing such as arrangement and mixing In these applications, music can be divided into several parts that are recombined based on special knowledge of musical composition to create new music. Recurrent neural network-based music composition systems are being developed to help composers to quickly compose music. Continuous recurrent neural networks based on the GAN (C-RNN-GAN) compose music by building a GAN model with two LSTMs, composing the melodies [11]. This study proposes a method for composing melody based on an enhanced GAN model.

Deep-Learning-Based Music Composition Methods
Comparison of Deep Learning-Based Music Generation Methods
Method
The generator
The Two Discriminators
Model Training
Experimental Environment
Experimental Data
Experimental Results
Loss Analysis of the Enhanced GAN Model
According
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call