Abstract

Deep clustering combines embedding and clustering together to obtain an optimal low dimensional embedding sub- space (aka latent subspace) for clustering, which can be more effective compared to conventional clustering approaches such as k-means. Typical deep clustering methods employ autoencoder (AE) and obtain their optimal latent space through minimizing data reconstruction loss which has no substantial connection with the clustering performance. In contrast, in this paper we propose a novel AE-based clustering scheme Deep Successive Subspace Learning (DSSL) which simultaneously minimizes weighted reconstruction and clustering losses of data points, where weights are defined based on similarity between latent representation of data points and cluster centers. DSSL obtains its optimal latent space through K (i.e. number of clusters) successive training runs where each run corresponds to an individual cluster. At each run, DSSL focuses on reconstruction and clustering of those data points that are more likely to belong to the corresponding cluster; hence, implicitly training those network parameters that have more influence on that cluster. Experimental results on benchmark datasets demonstrate that the proposed DSSL method can significantly outperform state-of-the-art clustering approaches.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.