Abstract

Vehicle re-identification (ReID) tasks are an important part of smart cities and are widely used in public security. It is extremely challenging because vehicles with different identities are generated from a uniform pipeline and cannot be distinguished based only on the subtle differences in their characteristics. To enhance the network’s ability to handle the diversity of samples in order to adapt to the changing external environment, we propose a novel data augmentation method to improve its performance. Our deep learning framework mainly consists of a local blur transformation and a transformation adversarial module. In particular, we first use a random selection algorithm to find a local region of interest in an image sample. Then, the parameter generator network, a lightweight convolutional neural network, is responsible for generating four weights and then as a basis to form a filter matrix for further blur transformations. Finally, an adversarial module is employed to ensure that as much noise information as possible is added to the image sample while preserving the structure of the training datasets. Furthermore, by updating the parameter generator network, the adversarial module can help produce more appropriate and harder training samples and lead to improving the framework’s performance. Extensive experiments on datasets, i.e., VeRi-776, VehicleID, and VERI-Wild, show that our method is superior to the state-of-the-art methods.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.