Abstract
BackgroundExploring correspondences across multiview images is the basis of various computer vision tasks. However, most existing methods have limited accuracy under challenging conditions. MethodTo learn more robust and accurate correspondences, we propose DSD-MatchingNet for local feature matching in this study. First, we develop a deformable feature extraction module to obtain multilevel feature maps, which harvest contextual information from dynamic receptive fields. The dynamic receptive fields provided by the deformable convolution network ensure that our method obtains dense and robust correspondence. Second, we utilize sparse-to-dense matching with symmetry of correspondence to implement accurate pixel-level matching, which enables our method to produce more accurate correspondences. ResultExperiments show that our proposed DSD-MatchingNet achieves a better performance on the image matching benchmark, as well as on the visual localization benchmark. Specifically, our method achieved 91.3% mean matching accuracy on the HPatches dataset and 99.3% visual localization recalls on the Aachen Day-Night dataset.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.