Abstract

Currently, monocular SLAM (simultaneous location and mapping) systems cannot extract depth information from monocular cameras directly and require initialization to solve the problem of scale uncertainty. It is extremely difficult to reconstruct maps using such systems, and it is difficult to cope with scenarios that require navigation and obstacle avoidance. In order to solve the above problems, in this paper, a simple monocular depth estimation network framework is proposed. Transfer learning from a pre-trained ResNet is utilized for the encoding part of the framework and a CNN is used as the decoding part. Only a few training parameters and iterations are required to obtain fairly accurate depth information. At the same time, a similarity-based filter is used to denoise the surfels and improve the RGB-D SLAM system, which not only reduces the impact of the depth estimation error on the surfels, but also ensures the quality of the dense mapping. From the results of comparative experiments, it can be seen that the proposed monocular depth estimation network framework is better than current popular methods, and the associated SLAM system can achieve pose estimation and dense mapping tasks. As a monocular camera-based SLAM system, the proposed method is a promising and practical approach.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.