Abstract

When handling large-scale multimodal data, hashing-based retrieval methods have gained significant attention due to their advantages in terms of low storage consumption and quick query speed. Despite the good performance of supervised hashing techniques, there are still certain limitations that need to be addressed. Firstly, they ignore valuable information such as reconstruction residuals when employing matrix factorization to learn the low-dimensional space. Secondly, most methods map heterogeneous data to the same feature space, which is unreasonable because different modalities possess distinct dimensions and distributions. Lastly, some methods relax constraints in order to optimize discrete hash codes, leading to considerable quantization errors. To solve the above problems, a novel two-step hashing-based technology called Symmetric Consistent Hashing (SCH) is proposed. Specifically, in the first step, two separate latent semantic spaces are learned by leveraging Principal Component Analysis to mitigate principal energy loss and Locality Preserving Projections to preserve local similarity. Then, the latent spaces of different modalities are aligned to facilitate the sharing of high-level semantic information by exploiting modality consistency. Next, hash codes are learned from distinct latent spaces, which are embedded in the second step for learning hash functions through Kernel Logistic Regression. To validate the effectiveness of SCH, extensive experiments on three publicly available benchmark datasets are accomplished. The results demonstrate the superiority of SCH over state-of-the-art hashing baselines, thus confirming the potency of the proposed approach.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call