Abstract

The paper is dedicated to the development and comparative experimental analysis of semi-supervised learning approaches based on a mix of unsupervised and supervised approaches for the classification of datasets with a small amount of labeled data, namely, identifying to which of a set of categories a new observation belongs using a training set of data containing observations whose category membership is known. Semi-supervised learning is an approach to machine learning that combines a small amount of labeled data with a large amount of unlabeled data during training. Unlabeled data, when used in combination with a small quantity of labeled data, can produce significant improvement in learning accuracy. The goal is semi-supervised methods development and analysis along with comparing their accuracy and robustness on different synthetics datasets. The proposed approach is based on the unsupervised K-medoids methods, also known as the Partitioning Around Medoid algorithm, however, unlike Kmedoids the proposed algorithm first calculates medoids using only labeled data and next process unlabeled classes – assign labels of nearest medoid. Another proposed approach is the mix of the supervised method of K-nearest neighbor and unsupervised K-Means. Thus, the proposed learning algorithm uses information about both the nearest points and classes centers of mass. The methods have been implemented using Python programming language and experimentally investigated for solving classification problems using datasets with different distribution and spatial characteristics. Datasets were generated using the scikit-learn library. Was compared the developed approaches to find average accuracy on all these datasets. It was shown, that even small amounts of labeled data allow us to use semi-supervised learning, and proposed modifications ensure to improve accuracy and algorithm performance, which was demonstrated during experiments. And with the increase of available label information accuracy of the algorithms grows up. Thus, the developed algorithms are using a distance metric that considers available label information.
 Keywords: Unsupervised learning, supervised learning. semi-supervised learning, clustering, distance, distance function, nearest neighbor, medoid, center of mass.

Highlights

  • A large amount of data was produced recently, and nowadays humanity had the opportunity to store and process all this data

  • The most efficient approach in machine learning is supervised learning, when we have some data with labels and try to learn a function on data points as label pairs

  • Unlike K-medoids the proposed algorithm first calculates medoids using only labeled data and process unlabeled classes – assign labels of nearest medoid. This algorithm has the following pros: – reduced processing time, because required only multiple iteration throw points unlike standard K-medoid; – more robustness to wrong assigned labels, because the algorithm gives higher weights to labeled data in the medoids calculation step. Another proposed approach uses the idea of K-nearest neighbors and K-Mean algorithm, because for classifying we use both information about the nearest points and classes centers of mass

Read more

Summary

Introduction

A large amount of data was produced recently, and nowadays humanity had the opportunity to store and process all this data. This algorithm has the following pros: – reduced processing time, because required only multiple iteration throw points unlike standard K-medoid; – more robustness to wrong assigned labels, because the algorithm gives higher weights to labeled data in the medoids calculation step Another proposed approach uses the idea of K-nearest neighbors and K-Mean algorithm, because for classifying we use both information about the nearest points and classes centers of mass (algorithm 1). The K-nearest neighbors based approach has better accuracy in case of closely located clusters with the same distribution Another required feature of a semi-supervised algorithm is quality versus a number of labels dependency: more labels – higher quality and vice versa.

Moons Aniso Varied accuracy
Conclusions
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.