Abstract

Musical notation is one thing that needs to be learned to play music. This notation has an important role in music because it can help in visualizing instructions for playing musical instruments and singing. Unfortunately, musical symbols that are commonly written in musical notation are difficult for beginners who have just started learning music. This research proposed a solution to create an optical music recognition (OMR) using a deep learning model to classify musical notes more accurately with some of the latest convolutional neural network (CNN) architectures. The research was carried out by implementing vision transformer (ViT), CoAtNet-0, and ConvNeXt-Tiny architecture. The training process was also combined with data augmentation to provide more information for the model to learn. Then the accuracy results of each model were compared to find out the best model for the OMR solution in this research. This experiment uses the Andrea dataset and Attwenger dataset which both get the best result by using the augmentation method and ConvNeXt-Tiny as the model. The best accuracy for the Andrea dataset is 98.15% and for the Attwenger dataset is 98.43%.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call