Abstract

RGB-Infrared cross-modality person re-identification is an important task for 24hour full-time intelligent video surveillance, the task is challenging because of cross modal heterogeneity and intra modal variation. A novel deep attention network is proposed in this paper to handle these challenges by increasing the discriminability of the learned person features. The method includes three elements: (1) dual-path CNN to extract the feature maps of the RGB images and infrared images respectively, (2) dual-attention mechanism combining spatial attention and channel attention to enhance the discriminability of extracted features, and (3) joint loss function joining bi-directional ranking loss and identity loss to constraint the training process to further increase the accuracy. Extensive experiments on two public datasets demonstrate the effectiveness of our proposed method because the method achieves higher performance than state-of-the-arts methods.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call