Abstract

Labeling large amount of data is one of important issues in deep learning due to high labeling cost. One method to address this issue is the use of active learning. Active learning selects from a large unlabeled data pool, a set of data that is more informative to training a model for the task at hand. Many active learning approaches use uncertainty-based methods or diversity-based methods. Both have had good results. However, using uncertainty-based methods, there is a risk that sampled data may be redundant, and the use of redundant data can adversely affect the training of the model. Diversity-based methods risk losing some data important for training the model. In this paper, we propose the uncertainty-based Selective Clustering for Active Learning (SCAL), a method of selectively clustering for data with high uncertainty to sample data from each cluster to reduce redundancy. SCAL is expected to extend the area of the decision boundary represented by the sampled data. SCAL achieves a cutting-edge performance for classification tasks on balanced and unbalanced image datasets as well as semantic segmentation tasks.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call