Abstract

As CNNs have a strong capacity to learn discriminative facial features, CNNs have greatly promoted the development of face recognition, where the loss function plays a key role in this process. Nonetheless, most of the existing loss functions do not simultaneously apply weight normalization, apply feature normalization and follow the two goals of enhancing the discriminative capacity (optimizing intra-class/inter-class variance). In addition, they are updated by only considering the feedback information of each mini-batch, but ignore the information from the entire training set. This paper presents a new loss function called Gico loss. The deep model trained with Gico loss in this paper is then called GicoFace. Gico loss satisfies the four aforementioned key points, and is calculated with the global information extracted from the entire training set. The experiments are carried out on five benchmark datasets including LFW, SLLFW, YTF, MegaFace and FaceScrub. Experimental results confirm the efficacy of the proposed method and show the state-of-the-art performance of the method.

Highlights

  • CNNs have greatly promoted the development of face recognition, where the loss function plays a key role in training the CNNs

  • This paper presents a new loss function, which is called Global Information-based Cosine Optimal loss (i.e., Gico loss), and the deep model trained with Gico loss is named GicoFace

  • After reviewing the recent loss functions used in deep face recognition, we present a new loss function, namely Gico loss (Global Information-based Cosine Optimal loss)

Read more

Summary

Introduction

CNNs have greatly promoted the development of face recognition, where the loss function plays a key role in training the CNNs. Many different loss functions [1,2,3,4,5,6,7,8,9,10,11,12] have been proposed for learning highly discriminative features for face recognition These loss functions can be broadly grouped into two categories—the Euclidean distance-based loss functions [1,2,3,4,5] and the cosine similaritybased loss functions [6,7,8,9,10,11,12], where the vast majority of these loss functions are derived from cross entropy loss by modifying cross entropy loss with additional constraints or adding a penalty to it. Only a few of them explicitly follow the aforementioned two targets

Methods
Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call