Abstract

Spectral clustering is a very popular graph-based clustering technique that partitions data groups based on the input data similarity matrix. Many past studies based on spectral clustering, however, do not consider the global discriminative structure of the dataset. Also, the benefits of using more than one kernel have not been fully exploited with respect to spectral clustering, although it has been established by these past studies that using more than one kernel in clustering can result in a more accurate clustering than those obtained with a single kernel. Multi-kernel approaches, however, tend to be more time consuming compared to single kernel methods. To compensate these drawbacks, we integrate a global discriminative term into the clustering with an adaptive neighbor framework. This is done to preserve both the global geometric information and global discriminative information in a dual kernel space, in an attempt to optimize clustering performance. Via co-regularization, we utilize more than one kernel space to take advantage of the benefits of multiple kernels. We, however, use two heterogeneous kernels to help us reduce clustering time, since the ability to quickly process data is as equally important as its accuracy in this era of information explosion. Since these different kernel spaces admit the same underlying clustering of the data, we approach the problem looking for clustering consistent across the two kernel views. Hence we are able to detect the non-linear intrinsic geometrical information of the dataset. We perform clustering using the obtained indicator matrix from our modified Laplacian utilizing k-means. Our Experimental outcomes show that our approach gives satisfactory results in terms of accuracy and NMI, with time-to-cluster savings in comparison to other state-of-the-art clustering methods using both synthetic and public datasets.

Highlights

  • Clustering is a very useful procedure in the field of artificial intelligence

  • These methods include K-means, Normalized Cut, Ratio Cut, Self-Tuning Spectral Clustering (ST-SC) [53], Local density adaptive similarity measurement for spectral clustering (DA-SC) [54], spectral clustering based on k-nearest neighbour [55], Spectral clustering with adaptive similarity measure (ASM-SC) [56] and spectral clustering with adaptive similarity measure in kernel space (ASMK-SC) [28]

  • The proposed approach is tested on publicly available datasets. For these publicly available dataset, we compare our approach with the Robust Kernel K-means (RKKM) [57], Self-Tuning Spectral Clustering (ST-SC) [53], Clustering with Adaptive Neighbor (CAN) [10], Spectral clustering with adaptive similarity measure (ASM-SC ) [56], Spectral clustering with adaptive similarity measure in Kernel space (ASMK-SC) [28], Robust Graph learning from Noisy data (RGC) [31], Low-rank Kernel Learning for Graph-based Clustering (LRKL) [29] and Clustering with Similarity Preserving (SPC and mSPC) [30]

Read more

Summary

INTRODUCTION

Clustering is a very useful procedure in the field of artificial intelligence. Based on clustering results, many analytic approaches could be carried out. 2) It uses a co-regularization approach to combine objectives of the individual kernel spaces with their disagreements to obtain a joint minimization problem that is solved to obtain a class indicator matrix used in k-means for clustering. PROPOSED METHOD The offered algorithm for our technique, Co-regularized Discriminative Spectral Clustering with Adaptive Similarity Measure in Dual Kernel Space (CoRDiSC-ASMDKS) is comprehensively outlined in this segment. The disagreement between the clusterings is measured and combined with the discriminative spectral clustering with adaptive similarity measure objectives of the individual kernel spaces, to obtain a joint minimization problem. 2) DISCRIMINATIVE SPECTRAL CLUSTERING WITH ADAPTIVE SIMILARITY MEASURE IN KERNEL SPACE A discrimination term ( ) is introduced into problem (5) for each kernel representation. Combining problem (21) with the discriminative spectral clustering with adaptive similarity measure objectives of individual kernel spaces, the following joint minimization problem is obtained. O(h3) + O(nh) + O(nk) + O(h − c) ∗ A where A is the number of restarted Arnoldi, h > c is the Arnoldi length used to compute the first c eigenvectors of our modified Laplacian matrix

EXPERIMENTS
PARAMETER SELECTION
ROBUSTNESS TO NOISY DATA
Findings
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call