Abstract

Semi-supervised learning is dedicated to solving the problem of poor model performance caused by the scarcity of labeled samples. Co-training algorithm, as a representative of semi-supervised learning algorithms, through constructing the diversity of classifiers, is used to solve the shortage of label samples that leads to low classifier accuracy. However, traditional Co-training style algorithms have strong constraints: (1) there are multiple views of data, and (2) the randomness of data views is relatively large, which leads to the low stability of the Co-training style algorithms. To optimize and solve the shortcomings of traditional Co-training style algorithms, a novel Co-training method based on sub-Kmeans named Op-FSCO is proposed. The proposed approach uses an optimal subspace construct method to extend Co-training to single-view data, greatly improving the application field of Co-training style algorithm. The clustering subspace is generated by plotting the probability of the correlation between the feature and the mutual information measure between the feature and the class label. For the resulting cluster subspace, the top two are selected to construct different views. Finally, experiments on a synthetic dataset, UCI dataset, and real-world dataset demonstrate that the proposed approach is more effective than other Co-training style algorithms. On the UCI benchmark dataset, the performance of the Op-FSCO algorithm on 15 UCI datasets is significantly better than the traditional Co-training algorithm. And on the cere bral stroke dataset, our approach is 2%–5% higher than that of traditional Co-training style algorithms.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.