Abstract

Generalized Zero-Shot Learning (GZSL) has become an important research due to its powerful ability of recognizing unseen objects. Generative methods, converting conventional GZSL to fully supervised learning, can achieve competing performance, and most of them use semantic attributes plus Gaussian noise to enrich generated features. The visual features obtained in this way are consistent with the semantic description. However, the reality is that the semantic description of the visual features of the same category should be different, because the appearance of images differs from each other although they belong to a same category, i.e., mapping from semantic attributes to visual features should be a “many to many” relationship rather than “one to many”. Therefore, we propose a novel method to generate diverse augmented attribute, which are subsequently utilized to synthesize features. We construct a semantic generator based on a pre-trained semantic mapper, which augments the category semantics. Using the augmented category semantics to generate visual features will result in a better fit of the generated visual features to the distribution of real features. The proposed method can well solve the pseudo diversity of visual features generated by most generative GZSL methods. We evaluate the proposed method on five popular benchmark datasets, and the results show that it can achieve the state-of-the-art performance.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call