Abstract

In classification tasks, training labels are usually specified as one-hot targets which represent each class equally and exclusively. However, this labeling rule is not suitable in some situations. For the dependent classes, one-hot targets are not capable to represent the relation among them. The existing label smoothing methods just split the target response into neighboring classes, but it is only applied for ordinal classification, but not for the dependent but non-ordered classes. In this paper, we propose a novel labeling rule that decomposes the one-hot target into several bases to reflect relationships among classes while maintaining a balanced target space, which adopts channel encoding from communication systems, in particular, Bose–Chaudhuri–Hocquenghem (BCH) encoding. Besides, BCH encoder has an error-correcting mechanism that is expected to lift the accuracy. In theory, training with BCH targets ensures improved classification performance given the original accuracy is not less than 50%. To verify the proposed method on dependent classification, we conduct experiments with two facial tasks: age recognition and face anti-spoofing. The former is an ordinal classification task, and the latter is also regarded as a specific dependent classification problem due to the varying attack types being classified as one class finally and real for the other. Experimental results show that the proposed method improves accuracy by 6.33% on age recognition and reduces HTER by 3.63% for face anti-spoofing. In addition, as BCH targets divide the original response into a higher dimensional space, that is, the model is made to be heeded on learning the delicate sub-features. Hence, BCH targets also enhance model generalizability, thus guaranteeing improved performance on cross-domain evaluations. We further perform an assessment on the PACS dataset for evaluating domain generalizability. The results show that the domain generalizability is enhanced by increasing average accuracy by over 2% training with BCH targets.

Highlights

  • Image classification commonly involves taking a set of onehot targets as training labels; the output feature is compared to each target, in terms of cosine similarity, after which the results are mapped to the corresponding class probabilities

  • To determine the theoretical improvements attainable when training with BCH targets, we first assume that the output neurons right after the trainable weights have the same probabilities and that the central features trained by the one-hot targets are evenly distributed in the feature space

  • Age recognition illustrates how the proposed method works on tasks with dependent and ordered classes; face anti-spoofing is for dependent-only classes

Read more

Summary

Introduction

Image classification commonly involves taking a set of onehot targets as training labels; the output feature is compared to each target, in terms of cosine similarity, after which the results are mapped to the corresponding class probabilities. Some soft labeling has been proposed to address this shortcoming: it distributes the target probability among neighboring classes with certain functions to satisfy the class characteristics and reflect relations among classes in a detailed manner. This label smoothing rule ensures that training targets represent gradual changes in conditions [1], [2]. Due to the varying types of attacks are all belong to the fake class and the other class is the real one Another example is binary facial emotion classification, which distinguishes the facial emotion into positive and negative sides

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call