Abstract

Deep learning-based models generalize better to unknown data samples after being guided “where to look” by incorporating human perception into training strategies. We made an observation that the entropy of the model’s salience trained in that way is lower when compared to salience entropy computed for models training without human perceptual intelligence. The research problem addressed by this paper is whether lowering the entropy of model’s class activation map helps in further increasing the performance, on top of the performance increase we observe for human saliency-based model’s training. In this paper we propose and evaluate four new entropy-based loss functions controlling the model’s focus, covering the full range of the level of such control, from none to its “aggresive” minimization. We show, using a problem of synthetic face detection, that improving the model’s focus, through lowering entropy by the proposed loss components, leads to models that perform better in an open-set scenario (in which the test samples are synthesized by unknown generative models): the obtained average Area Under the ROC curve (AUROC) ranges from 0.72 to 0.78, compared to AUROC = 0.64 observed for a state-of-the-art human-salience-only-based control of the model’s focus. We also show that optimal performance is obtained when the model’s loss function blends three aspects: regular classification performance, low-entropy of the model’s focus, and closeness of the model’s focus to human saliency. The major conclusion from this work is that maximization of the model’s focus is an important regularizer allowing the models to generalize better in an open set scenario. Future work directions include methods of blending classification-, human salience-, and model’s salience entropy-based loss components to achieve optimal performance in other domains than the synthetic face detection.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.