Abstract

Multiview clustering seeks to partition objects via leveraging cross-view relations to provide a comprehensive description of the same objects. Most existing methods assume that different views are linear transformable or merely sampling from a common latent space. Such rigid assumptions betray reality, thus leading to unsatisfactory performance. To tackle the issue, we propose to learn both common and specific sampling spaces for each view to fully exploit their collaborative representations. The common space corresponds to the universal self-representation basis for all views, while the specific spaces are the view-specific basis accordingly. An iterative self-supervision scheme is conducted to strengthen the learned affinity matrix. The clustering is modeled by a convex optimization. We first solve its linear formulation by the popular scheme. Then, we employ the deep autoencoder structure to exploit its deep nonlinear formulation. The extensive experimental results on six real-world datasets demonstrate that the proposed model achieves uniform superiority over the benchmark methods.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call