Abstract

Many real-world datasets are represented by multiple features or modalities which often provide compatible and complementary information to each other. In order to obtain a good data representation that synthesizes multiple features, researchers have proposed different multi-view subspace learning algorithms. Although label information has been exploited for guiding multi-view subspace learning, previous approaches did not well capture the underlying semantic structure in data. In this paper, we propose a new multi-view subspace learning algorithm called multi-view semantic learning (MvSL). MvSL learns a nonnegative latent space and tries to capture the semantic structure of data by a novel graph embedding framework, where an affinity graph characterizing intra-class compactness and a penalty graph characterizing inter-class separability are generally defined. The intuition is to let intra-class items be near each other while keeping inter-class items away from each other in the learned common subspace across multiple views. We explore three specific definitions of the graphs and compare them analytically and empirically. To properly assess nearest neighbors in the multi-view context, we develop a multiple kernel learning method for obtaining an optimal kernel combination from multiple features. In addition, we encourage each latent dimension to be associated with a subset of views via sparseness constraints. In this way, MvSL is able to capture flexible conceptual patterns hidden in multi-view features. Experiments on three real-world datasets demonstrate the effectiveness of MvSL.

Highlights

  • In many real-world data analytic problems, instances are often described with multiple modalities or views

  • Since the large-margin based multi-view learning methods ignore the intra-class semantic structures of the data, the latent subspace learned by this kind of methods will acquire inferior representation compared with MvFisher and multi-view semantic learning (MvSL)

  • We have proposed a novel nonnegative latent representation learning algorithm, called Multi-view semantic learning (MvSL), for representation learning with multi-view data

Read more

Summary

Introduction

In many real-world data analytic problems, instances (items) are often described with multiple modalities or views. In terms of the style of incorporating label information, existing supervised or semi-supervised multi-view representation learning methods can be divided into three categories: (1) large-margin based methods [6, 9, 15] This kind of methods uses the large-margin principle to maximize the margin between instances of different classes, but ignores the intra-class semantic structures. Fisher’s discriminant analysis based methods are optimal only in cases where the data of each class follows Gaussian distribution In reality, this assumption is too restrictive since real world datasets often exhibit complex non-Gaussian distributions [20, 21]. A sub-challenge in LDGE and TGE is how to identify nearest neighbors in the multi-view context To this end, we develop a new multiple kernel learning algorithm to find the optimal kernel combination for multi-view features. The encouraging results of MvSL are achieved in comparison with the state-of-the-art algorithms

Label exploitation in multi‐view subspace learning
NMF and multi‐view extensions
Graph embedding
Multi‐view semantic learning
Multi‐view NMF
Graph embedding framework
Sparseness constraint
Graph embedding for multi‐view semantic learning
Simple graph embedding
Transductive graph embedding
Local discriminant graph embedding
Multiple kernel learning
Optimization
Optimizing
Optimizing MvSL‐S and MvSL‐L
Computational complexity
Data set
Evaluation methodology
Experiment results
Parameter sensitive analysis
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call