Abstract

The massive number of images demands highly efficient image retrieval tools. Deep distance metric learning (DDML) is proposed to learn image similarity metrics in an end-to-end manner based on the convolution neural network, which has achieved encouraging results. The loss function is crucial in DDML frameworks. However, we found limitations to this model. When learning the similarity of positive and negative examples, the current methods aim to pull positive pairs as close as possible and separate negative pairs into equal distances in the embedding space. Consequently, the data distribution might be omitted. In this work, we focus on the distribution structure learning loss (DSLL) algorithm that aims to preserve the geometric information of images. To achieve this, we firstly propose a metric distance learning for highly matching figures to preserve the similarity structure inside it. Second, we introduce an entropy weight-based structural distribution to set the weight of the representative negative samples. Third, we incorporate their weights into the process of learning to rank. So, the negative samples can preserve the consistency of their structural distribution. Generally, we display comprehensive experimental results drawing on three popular landmark building datasets and demonstrate that our method achieves state-of-the-art performance.

Highlights

  • Along with the popularity of the Internet and smart devices, the number of pictures stored on the network has exploded

  • The region descriptor is processed by L2 normalized, through principal component analysis (PCA) + whitened [32], and normalized by L2 and aggregated into a global signature called regional maximum activations of convolutions (RMAC) [33]

  • In order to be ablethe to formula highlightofthis effect of training,of wethe propose ansample entropyand weight based on structural distribution, which is in the process of training, we propose an entropy weight based on structural distribution, the formula shown in Equation (9)

Read more

Summary

Introduction

Along with the popularity of the Internet and smart devices, the number of pictures stored on the network has exploded. For a known query image, only a small amount of data is combined to train the network and calculate the loss, and pull the samples with low similarity to the same distance from the query image, ignoring some useful examples and structural distribution information [23]. For a positive sample, we force the distance between the positive sample and the query image to be less than the threshold, rather than pulling the positive sample more and more compact In this case, the DSLL can help preserve the similarity structure of each sample and the ordinal relationship of the data as much as possible. The algorithm considers the distribution characteristics of the negative samples, captures the feature structure, addresses the relative similarity of each sample, which has been ignored in past works, and calculates the loss by using weight to rank the negative samples.

Image Retrieval
Deep Metric Learning
Contrastive Loss
Triplet Loss
N-Pair Loss
Lifted Structured Loss
Proxy-NCA
Method
CNN Network Architecture
The Architecture of Network Training
Network Evaluting Architecture
Distributed
The sample andthat thewhen positive are represented
The Process of Distributed Structure Learning Loss
Experiments
Training Datasets
Training Configurations
Test Datasets
Performance Evaluation Metrics
The Impact of Margin Parameter τ
The Role of Structural Similarity Ranking Consistency
Evaluation is based performance with
The Combined Impact of DSLL
Evaluation is is
Comparison with the State of the Art
Visualization Purposes
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call