Abstract

Visual Simultaneous Localization and Mapping (VSLAM) is a prerequisite for robots to accomplish fully autonomous movement and exploration in unknown environments. At present, many impressive VSLAM systems have emerged, but most of them rely on the static world assumption, which limits their application in real dynamic scenarios. To improve the robustness and efficiency of the system in dynamic environments, this paper proposes a dynamic RGBD SLAM based on a combination of geometric and semantic information (DGS-SLAM). First, a dynamic object detection module based on the multinomial residual model is proposed, which executes the motion segmentation of the scene by combining the motion residual information of adjacent frames and the potential motion information of the semantic segmentation module. Second, a camera pose tracking strategy using feature point classification results is designed to achieve robust system tracking. Finally, according to the results of dynamic segmentation and camera tracking, a semantic segmentation module based on a semantic frame selection strategy is designed for extracting potential moving targets in the scene. Extensive evaluation in public TUM and Bonn datasets demonstrates that DGS-SLAM has higher robustness and speed than state-of-the-art dynamic RGB-D SLAM systems in dynamic scenes.

Highlights

  • Simultaneous localization and mapping (SLAM) [1] is a crucial technology to achieve the autonomous perception of intelligent robots for applications such as augmented reality, indoor navigation, and autonomous vehicles [2,3,4]

  • We extend the work of ORB-SLAM3 and propose a fast and robust dynamic RGB-D SLAM system (DGS-SLAM) based on the combination of geometric models and semantic information

  • An overview of DGS-SLAM, which is implemented based on the RGB-D mode of the potential motion information of the historical semantic frames

Read more

Summary

Introduction

Simultaneous localization and mapping (SLAM) [1] is a crucial technology to achieve the autonomous perception of intelligent robots for applications such as augmented reality, indoor navigation, and autonomous vehicles [2,3,4] It enables the robot’s positional estimation and scene construction in unknown environments by analyzing its onboard sensors. The environment is static, and the change of field of view only comes from the camera’s motion, which limits its applicability in the real world Dynamic objects such as moving people, animals, and vehicles will lead to many incorrect data associations, which will further have severe negative impacts on pose estimation and scene construction. The effect of dynamic feature points can be reduced using methods such as bundle adjustment (BA) [8] or Random

Methods
Results
Discussion
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.