Abstract

Deep convolutional neural networks have achieved tremendous success in a variety of applications across many disciplines. However, their superior performance relies on correctly annotated large-scale datasets. It is very expensive and time-consuming to get the annotated large-scale datasets, especially in the medical field. While collecting a large amount of data is relatively easy, given the amount of data available on the web, but these data are highly unreliable, and they often include a massive amount of noisy labels. The past research works have shown that these noisy labels could significantly affect the performance of the deep convolutional neural networks on image classification. However, training a robust deep convolutional neural network with extremely noisy labels is a very challenging task. Inspired by the co-teaching concept, this paper proposes a novel method for training a robust convolutional neural network with extremely noisy labels, which is called group-teaching. Specifically, we train a group of convolutional neural networks simultaneously, and let them teach each other by selecting possibly clean samples for each network in each mini-batch. Each network back propagates the samples selected by other networks except itself and then it updates itself. The empirical results on noisy versions of CIFAR-10 and CIFAR-100 datasets demonstrate that our method is superior to the state-of-the-art methods in the robustness for noisy labels. Particularly, to verify the efficacy of our group-teaching in real-world noisy labels distribution, we have also validated the effectiveness of our method on the real-world noisy WebVision1000-100 dataset. The results show that our method has achieved higher performance than the state-of-the-art methods.

Highlights

  • In recent years, the deep convolutional neural networks have achieved tremendous success in numerous computer vision tasks, achieving state-of-the-art performance on image classification [1]–[6], object detection [7]–[10], semantic segmentation [11]–[14], and so on

  • In order to combat this drawback, we propose a novel method for training a robust convolutional neural network with extremely noisy labels, which is called group-teaching

  • EXPERIMENTAL RESULTS we empirically evaluate our method on the closed-set [42] noisy versions of CIFAR-10 [33] and CIFAR-100 [34] datasets, and the open-set [42] real-world noisy dataset WebVision1000-100, which is a 100-category subset of WebVision [18]

Read more

Summary

INTRODUCTION

The deep convolutional neural networks have achieved tremendous success in numerous computer vision tasks, achieving state-of-the-art performance on image classification [1]–[6], object detection [7]–[10], semantic segmentation [11]–[14], and so on. The authors in [26]–[28] introduced the techniques to down-weight the noisy samples by re-weighting the training samples Both methods in [26] and [27] required a smaller clean dataset to work. Another method in [28] derived an optimal important weighting scheme for noise-robust classification Other methods, such as decoupling [29] and co-teaching [30], re-sampled the possibly clean samples from the noisy data and updated the networks using these selected clean data. In order to combat this drawback, we propose a novel method for training a robust convolutional neural network with extremely noisy labels, which is called group-teaching. Our empirical results on noisy versions of CIFAR-10, CIFAR-100 and WebVision1000-100 datasets demonstrate that our method is superior to the state-of-theart methods in the robustness for noisy labels

RELATED WORK
EXPERIMENTAL RESULTS
NETWORK ARCHITECTURE
Findings
CONCLUSION

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.