Abstract

The feature quality is critical for positioning accuracy of feature-descriptor-matching-based visual SLAM (Visual Simultaneous Localization and Mapping). State-of-the-art handcrafted descriptors like BRIEF and ORB can hardly work well in complex scenarios. On the other hand, CNN (Convolutional Neural Network) is proved to have tremendous advantages on image feature extraction. In this paper, we develop a CNN model to extract binary visual feature descriptor from image patches with four important loss functions, namely adaptive scale loss, even distribution loss, quantization loss and correlation loss. Based on this learned deep binary feature descriptor sharing the same structure with ORB descriptor, a monocular SLAM system named DBLD-SLAM is designed, by replacing the ORB descriptor in conventional ORB-SLAM. We also train visual Bag of Words to detect loop closure. Experiments show that our DBLD descriptor achieves better results on the HPatches dataset and UBC benchmark. Moreover, the DBLD-SLAM system outperforms other current popular SLAM system on Tartanair dataset.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.