Abstract

Face representation learning using datasets with a massive number of identities requires appropriate training methods. Softmax-based approach, currently the state-of-the-art in face recognition, in its usual "full softmax" form is not suitable for datasets with millions of persons. Several methods, based on the "sampled softmax" approach, were proposed to remove this limitation. These methods, however, have a set of disadvantages. One of them is a problem of "prototype obsolescence": classifier weights (prototypes) of the rarely sampled classes receive too scarce gradients and become outdated and detached from the current encoder state, resulting in incorrect training signals. This problem is especially serious in ultra-large-scale datasets. In this paper, we propose a novel face representation learning model called Prototype Memory, which alleviates this problem and allows training on a dataset of any size. Prototype Memory consists of the limited-size memory module for storing recent class prototypes and employs a set of algorithms to update it in appropriate way. New class prototypes are generated on the fly using exemplar embeddings in the current mini-batch. These prototypes are enqueued to the memory and used in a role of classifier weights for softmax classification-based training. To prevent obsolescence and keep the memory in close connection with the encoder, prototypes are regularly refreshed, and oldest ones are dequeued and disposed of. Prototype Memory is computationally efficient and independent of dataset size. It can be used with various loss functions, hard example mining algorithms and encoder architectures. We prove the effectiveness of the proposed model by extensive experiments on popular face recognition benchmarks.

Highlights

  • The best performing models in popular face recognition benchmarks are trained on a datasets, containing hundreds of thousands [5], [8], [20] or even millions [21], [22] of persons and up to hundreds of millions of face images

  • In this paper we proposed Prototype Memory - a novel face representation learning model, opening the possibility to train state-of-the-art face recognition architectures on the datasets of any size

  • Prototype Memory consists of the memory module for storing class prototypes, and algorithms to perform operations with it

Read more

Summary

INTRODUCTION

F ACE recognition is one of the most established technologies [1], [2] of computer vision. "Sampled softmax"-based methods are good solutions for some problems in face recognition model training on largescale datasets, but they have a set of disadvantages They still need to store prototypes of all classes in memory. PROPOSED METHOD we propose Prototype Memory model for face representation learning It combines the advantages of weight imprinting (online generation of class prototypes using groups of class exemplars), exemplar memory (keeping useful information between training iterations) and accelerated softmax methods (faster and more memory-efficient than "full softmax") and could be used to train face recognition architectures on datasets of any size without the problem of prototype obsolescence. New class prototypes are enqueued to the Prototype Memory module Prototypes in this module are later used as the classifier weights to perform usual softmax-based face recognition model training with some appropriate loss function. Since new class prototypes are generated using up-to-date encoder, the distance between class centers and prototypes stays small, so the problem of prototype obsolescence is solved

TRAINING SPECIFICS
PROTOTYPE MEMORY KNOWLEDGE DISTILLATION
EXPERIMENTS
HYPERPARAMETER EFFECTS We have performed the experiments with different hyperparameters of Prototype Memory
EXPERIMENTS WITH PROTOTYPE OBSOLESCENCE
COMPARISON WITH OTHER "SAMPLED SOFTMAX"-BASED METHODS
Method
EXPERIMENTS WITH MULTI-DOPPELGANGER MINING AND HARDNESS-AWARE EXAMPLE MINING
EXPERIMENTS WITH KNOWLEDGE DISTILLATION
CONCLUSION
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.