Abstract
Early childhood children frequently use body movements while singing songs accompanied to the music. When trying to express the recognition of musical elements by body movement, the child in early childhood creates his own musical expression by thinking and judging while looking at the surrounding children and the accompaniment teacher. This study aims to quantitatively analyze changes in eye movements during musical expression. 3-year-old, 4-year-old, and 5-year-old children at two nursery schools in 2020 and two kindergartens in 2021 (n=118) participated in eye tracking during singing a song using an eye tracker (Tobii3). Quantitative analysis by three-way ANOVA was mainly conducted on the calculated data. As a result, the increase in data such as number of saccade occurrences and size, and the moving average velocity of saccade, showed that saccades during musical expression in early childhood tended to be larger in major key than in minor key. From the calculated data on saccade, which is the eye movement during musical expression, it was predicted that effective feature quantities of eye movement during musical expression for machine learning could be derived in the same way as feature quantities depended on the results of quantitative analysis of body movement during musical expression.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.