Abstract

Positive and Unlabeled learning (PU learning) trains a binary classifier based on only positive (P) and unlabeled (U) data, where the unlabeled data contains positive or negative samples. Previous importance reweighting approaches treat all unlabeled samples as weighted negative samples, achieving state-of-the-art performance. However, in this paper, we surprisingly find that the classifier could misclassify negative samples in U data as positive ones at the late training stage by weight adjustment. Motivated by this discovery, we leverage Semi-Supervised Learning (SSL) to address this performance degradation problem. To this end, we propose a novel SSL-based framework to tackle PU learning. Firstly, we introduce the dynamic increasing sampling strategy to progressively select both negative and positive samples from U data. Secondly, we adopt MixMatch to take full advantage of the unchosen samples in U data. Finally, we propose the Co-learning strategy that iteratively trains two independent networks with the selected samples to avoid the confirmation bias. Experimental results on four benchmark datasets demonstrate the effectiveness and superiority of our approach when compared with other state-of-the-art methods.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call