Abstract

Person reidentification involves recognizing a person across nonoverlapping camera views, with different pose, illumination, and camera characteristics. We propose to tackle this problem by training a deep convolutional network to represent a person’s appearance as a low-dimensional feature vector that is invariant to common appearance variations encountered in the reidentification problem. Specifically, a Siamese network architecture is used to train a feature extraction network using pairs of similar and dissimilar images. We show that the use of a novel multitask learning objective is crucial for regularizing the network parameters in order to prevent overfitting due to the small size of the training data set. We complement the verification task, which is at the heart of reidentification, by training the network to jointly perform verification and identification and to recognize attributes related to the clothing and pose of the person in each image. In addition, we show that our proposed approach performs well even in the challenging cross-data set scenario, which may better reflect real-world expected performance.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.