Abstract
Compared with conventional ReID task, RGB-IR person re-identification is more significant and challenging owing to enormous cross-modality variations between RGB and infrared images. Most existing cross-modality person re-identification approaches in the infrared and visible domains use the joint training of traditional triplet loss and CE loss to reduce the discrepancy between two modalities. However, as the model gradually converges, the number of positive and negative sample pairs that can be optimized by traditional triplet loss also decreases. In this paper, we adopt a dual-stream CNN structure and propose a maximum intra-class distance to triplet loss (MICT loss) for RGB-IR ReID task. The proposed method enjoys several advantages. First, it can learn specific modal information and shared information better through the CNN structure of the dual-stream branch. Second, it can add the optimization of intra-class cross-modality distance to the traditional triplet loss, which can effectively reduce the intra-class distance and increase the training difficulty of the model. Experimental results demonstrate that our proposed algorithm performs favorably against the state-of-the-art methods that rank-1 accuracy is 61.71% and mAP is 58.06% on SYSU-MM01 dataset.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.