Abstract

As a prominent semi-supervised learning algorithm, co-training can make full use of a few labeled samples as well as a quantity of unlabeled samples to train robust classifiers, thus it is widely researched in the last decades. However, there are still two severe problems in existing co-training methods: firstly, data used in co-training needs at least two sufficient and redundant views, but few practical datasets can meet it. Secondly, because there are a few labeled samples in the initial stage, the initial classifiers trained by co-training are usually too weak to correctly label unlabeled samples, which may bring in noisy labels for the following training process. To deal with these problems, we propose a co-training algorithm by predicting confidence of unlabeled neighbors (termed LCN-CoTrain). Specifically, LCN-CoTrain first defines an entropy-based view division method to generate redundant views. Meanwhile, LCN-CoTrain introduces a novel labeling confidence prediction strategy, which takes the nearest unlabeled neighbors into consideration when calculating the labeling confidence of a certain sample. To validate the efficiency of our proposed LCN-CoTrain algorithm, we experiment on four UCI datasets and compare LCN-CoTrain with several representative co-training methods. The experimental results indicate that our proposed LCN-CoTrain can learn robust classifiers and outperform most of baseline methods.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.