Abstract

K-means is the most well-known and widely used classical clustering method, benefited from its efficiency and ease of implementation. But k-means has three main drawbacks: the selection of its initial cluster centers can greatly affect its final results, the number of clusters has to be predefined, and it can only find clusters of similar sizes. A lot of work has been done on improving the selection of the initial cluster centers and on determining the number of clusters. However, very little work has been done on improving k-means to deal with clusters of different sizes. In this paper, we have proposed a new clustering method, called k-normal, whose main idea is to learn cluster sizes during the same process of learning cluster centers. The proposed k-normal method can identify clusters of different sizes while keeping the efficiency of k-means. Although the Expectation Maximization (EM) method based on Gaussian mixture models can also identify the clusters of different sizes, it has a much higher computational complexity than both k-normal and k-means. Experiments on a synthetic dataset and seven real datasets show that, k-normal can outperform k-means on all the datasets. If compared with the EM method, k-normal still produces better results on six out of the eight datasets while enjoys a much higher efficiency.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.