Abstract
Deep neural networks (DNNs) have achieved astonishing results on a variety of supervised learning tasks owing to a large scale of expert-labeled datasets. However, as recent researches pointed out, the generalization performance of DNNs deteriorates rapidly when training data contains label noise. To alleviate the problem of easily overfitting to label-corrupted data when training DNNs, this paper proposes a robust loss function by reweighting the standard Cross-Entropy loss. For obtaining more robust DNNs under label noise, we further design a framework to jointly optimize model parameters and filtering noisy data during training. Our methods can be thus viewed as online curriculum learning based on both loss function and training datasets. A great deal of experiments have conducted on CIFAR-10, CIFAR-100 and ImageNet under two types of label noise. The results of our proposed method outperforms other state-of-the-art methods.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.