Abstract
Solving the problem of pedestrians being occluded by objects is extremely challenging. Using part-level features to describe pedestrian images can provide fine-grained information. However, only paying attention to the local features of body will lack global pedestrian information. And the network consumes time and memory. To solve these problems, we propose a new person re-identification network. The network uses a global contrastivemodule to obtain the features of pedestrians. Through effective use of the pedestrian’s global features, as well as the pedestrian’s personal information and global contrastive information, the pedestrian can be found in the object occlusion to provide a reliable feature embedding. Our model is tested on Market1501, Duke MTMC-reID, CUHK03 and MSMT17 datasets. The experimental results show that our method is effective in occluded person re-identification
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.