Image aesthetics assessment (IAA) is inherently subjective, with generic image aesthetics assessment (GIAA) often overlooking individual user preferences. Personalized aesthetic assessment (PIAA) addresses this by accounting for user-specific inconsistencies. However, the limited annotated data from individual users and the scarcity of user-jointly annotated image data pose challenges. To cope with these problems, we propose a multi-scale regional attribute-weighting via meta-learning for PIAA, namely MRAM, which aims to simulate how the brain computes aesthetic values from visual stimuli. The human brain is to predict the subjective aesthetic value of the individual to the image by combining the features of the visual scene segmentation through the weighting method. MRAM introduces a multi-scale attention module to capture visual scene regions segmented at different scales, adjusting feature selection based on user attention. Additionally, a regional attribute-weighting module (RAM) is proposed to weigh these features using aesthetic attribute representations. In the meta-training phase, MRAM learns aesthetic prior knowledge from a diverse set of PIAA tasks. Transitioning to the personalization phase, MRAM is fine-tuned with only a small number of specific user samples, enabling it to rapidly adapt to the unique aesthetic preferences of individuals. Experimental results demonstrate the superior performance of the proposed MRAM compared to state-of-the-art PIAA methods.
Read full abstract