Abstract

Deep, convolutional neural networks have become the state-of-the-art method for automatic Facial Expression Recognition (FER). Because of the small size and controlled conditions of most FER datasets, however, models can still overfit to the training dataset and struggle to generalize well to new data. We present a novel approach of using crowdsourced label distributions for improving the generalization performance of convolutional neural networks for FER. We implement this as a loss layer regularizer, where the ground truth labels are combined with crowdsourced labels in order to construct a noisy output distribution during training. We use a label disturbance method in which training examples are randomly replaced with incorrect labels drawn from the combined label probability distribution. We compare the performance of our disturbed and undisturbed models in cross-validation testing on the extended Cohn-Kanade dataset and cross-dataset experiments on the MMI, JAFFE, and FER2013 datasets. We find that using our proposed method, test performance is improved on both the MMI and JAFFE datasets. Our results suggest that using nonuniform probability distributions to disturb training can improve generalization performance of CNNs on other FER datasets.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.