Abstract

Hashing offers a desirable and effective solution for efficiently retrieving the nearest neighbors from large-scale data because of its low storage and computation costs. One of the most appealing techniques for hashing learning is matrix factorization. However, most hashing methods focus only on building the mapping relationships between the Euclidean and Hamming spaces and, unfortunately, underestimate the naturally sparse structures of the data. In addition, parameter tuning is always a challenging and head-scratching problem for sparse hashing learning. To address these problems, in this article, we propose a novel hashing method termed adaptively sparse matrix factorization hashing (SMFH), which exploits sparse matrix factorization to explore the parsimonious structures of the data. Moreover, SMFH adopts an orthogonal transformation to minimize the quantization loss while deriving the binary codes. The most distinguished property of SMFH is that it is adaptive and parameter-free, that is, SMFH can automatically generate sparse representations and does not require human involvement to tune the regularization parameters for the sparse models. Empirical studies on four publicly available benchmark data sets show that the proposed method can achieve promising performance and is competitive with a variety of state-of-the-art hashing methods.

Highlights

  • W ITH the rapid advancement of information technologies and the prevalence of social networks, considerable quantities of data, including images, videos, and texts collected from different application domains, such as computer vision, machine learning, and information retrieval, are increasing in a marvelous and unprecedented way

  • We focus our attention on hashing learning and propose a novel two-stage hashing method, dubbed as adaptively sparse matrix factorization hashing (SMFH)

  • The comparison results of SMFH with the state-of-the-art hashing algorithms are given

Read more

Summary

INTRODUCTION

W ITH the rapid advancement of information technologies and the prevalence of social networks, considerable quantities of data, including images, videos, and texts collected from different application domains, such as computer vision, machine learning, and information retrieval, are increasing in a marvelous and unprecedented way. The data-dependent hashing ( known as learning-based hashing) seeks projection functions to capture the underlying geometries of data Typical examples of such methods are PCA hashing (PCAH) [17], spectral hashing (SH) [18], iterative quantization (ITQ) [19], circulant binary embedding (CBE) [20], density sensitive hashing (DSH) [21], spherical hashing (SpH) [22], and sparse embedding and least variance encoding (SELVE) [23]. 1) We formulate the learning method for hashing as an optimization problem and solve it by using matrix factorization, which is good at capturing the latent structural information of the data.

RELATED WORK
SPARSE MATRIX FACTORIZATION HASHING
Notations and Problem Statements
Adaptively Sparse Matrix Factorization
Algorithm Implementation
EVALUATION EXPERIMENTS
Data Sets and Settings
Baselines and Metrics
Results and Discussion
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call