Abstract

The performance of face recognition has been boosted by the features extracted from deep convolutional neural networks. Ideal features should have minimum intra-class variations and maximum inter-class variations. The most commonly used loss function for classification, softmax loss, however, does not necessarily learn features discriminative enough. Large margin classifiers have nice generalization properties in statistical machine learning. These properties have lead to the application of margin to deep learning in recent years. We hereby propose a new loss function called Contrapositive Margin Softmax loss for face verification task, which helps to learn invariant and discriminative features by introducing margins to both target logits and maximum negative logits of softmax loss. Competitive results on LFW (99.28%) and YTF (95.34%) demonstrate the effectiveness of our approach.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.