Abstract

Unsupervised person re-identification (Re-ID) aims at finding the most informative features from unlabeled person datasets. Some recent approaches adopted camera-aware strategies for model training and have thereby achieved highly promising results. However, these methods simultaneously address intra-ID discrepancies of all cameras and require independent learning under each camera, which increases the complexity of algorithm. To resolve this issue, we present a camera contrast learning framework for unsupervised person Re-ID. Our method first proposes a time-based camera contrastive learning module to facilitate model learning. At each iteration, we follow the time contrast principle to select one camera centroid as proxy of each cluster. By enforcing the samples to converge to positive proxies, the correlation between features and cameras can gradually be reduced. Moreover, we design a 3-dimensional attention module to further reduce intra-ID discrepancies caused by background shifts. By re-weighting each feature map element in a spatial-channel order, our module can exactly find identity-invariant semantic cues from regions of interest in person images, no matter how the background change. Experimental results on several popular datasets prove that our work surpasses existing unsupervised person Re-ID approaches to a remarkable extent. The source codes can be found in https://github.com/HongweiZhang97/CCL.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call