Abstract

Center-based clustering methods like k-Means intend to identify closely packed clusters of data points by respectively finding the centers of each cluster. However, k-Means requires the user to guess the number of clusters, instead of estimating the same on the run. Hence, the incorporation of accurate automatic estimation of the natural number of clusters present in a data set is important to make a clustering method truly unsupervised. For k-Means, the minimum of the pairwise distance between cluster centers decreases as the user-defined number of clusters increases. In this paper, we observe that the last significant reduction occurs just as the user-defined number surpasses the natural number of clusters. Based on this insight, we propose two techniques: the Last Leap (LL) and the Last Major Leap (LML) to estimate the number of clusters for k-Means. Over a number of challenging situations, we show that LL accurately identifies the number of well-separated clusters, whereas LML identifies the number of equal-sized clusters. Any disparity between the values of LL and LML can thus inform a user about the underlying cluster structures present in the data set. The proposed techniques are independent of the size of the data set, making them especially suitable for large data sets. Experiments show that LL and LML perform competitively with the best cluster number estimation techniques while imposing drastically lower computational burden.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call