Abstract

Generalized zero-shot learning suffers from an extreme data imbalance problem, that is, the training data only come from seen classes while no unseen class data are available. Recently, a number of feature generation methods based on generative adversarial networks (GAN) have been proposed to address this problem. Existing feature generation methods, however, have never considered the under-constrained problem, and thus could generate an unrestricted visual feature corresponding to no meaningful object class. In this paper, we propose to equip the feature generation framework with a parallel inference network that projects visual feature to the semantic descriptor space, constraining to avoid the generation of unrestricted visual features. The two-parallel-stream framework (1) enables our method, termed inference guided feature generation (Inf-FG), to mitigate the under-constrained problem and (2) makes our Inf-FG applicable to transductive ZSL. Our Inf-FG learns the feature generator and the inference network simultaneously by aligning the joint distribution of visual features and semantic descriptors from the feature generator and the joint distribution from the inference network. We evaluate our approach on four benchmark ZSL datasets, including AWA, CUB, SUN, and FLO, on which our method improves our baselines on generalized zero-shot learning.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call