Abstract

RGB-D SLAM (Simultaneous Localization and Mapping) generally performs smoothly in a static environment. However, in dynamic scenes, dynamic features often cause wrong data associations, which degrade accuracy and robustness. To address this problem, in this paper, a new RGB-D dynamic SLAM method, PLD-SLAM, which is based on point and line features for dynamic scenes, is proposed. First, to avoid under-over segmentation caused by deep learning, PLD-SLAM combines deep learning for semantic information segmentation with the K-Means clustering algorithm considering depth information to detect the underlying dynamic features. Next, two consistency check strategies are utilized to check and filter out the dynamic features more reasonably. Then, to obtain a better practical performance, point and line features are utilized to calculate camera pose in the dynamic SLAM, which is also different from most published dynamic SLAM algorithms based merely on point features. The optimization model with point and line features is constructed and utilized to calculate the camera pose with higher accuracy. Third, enough experiments on the public TUM RGB-D dataset and the real-world scenes are conducted to verify the location accuracy and performance of PLD-SLAM. We compare our experimental results with several state-of-the-art dynamic SLAM methods in terms of average localization errors and the visual difference between the estimation trajectories and the ground-truth trajectories. Through the comprehensive comparisons with these dynamic SLAM schemes, it can be fully demonstrated that PLD-SLAM can achieve comparable or better performances in dynamic scenes. Moreover, the feasibility of camera pose estimation based on both point features and line features has been proven by the corresponding experiments from a comparison with our proposed PLD-SLAM only based on point features.

Highlights

  • Simultaneous Localization and Mapping (SLAM) [1] was first proposed by researchers in 1986 and has been used to estimate the pose of a mobile robot and construct an incremental map of unknown surroundings simultaneously [2,3]

  • For the well-known SLAM systems (e.g., ORB-SLAM2 [10]), the RMSE is usually utilized to evaluate the overall performance of SLAM

  • We presented a new RGB-D SLAM based on point and line features in the dynamic scenes

Read more

Summary

Introduction

Simultaneous Localization and Mapping (SLAM) [1] was first proposed by researchers in 1986 and has been used to estimate the pose of a mobile robot and construct an incremental map of unknown surroundings simultaneously [2,3]. SLAM is the key technology for robots navigating in unknown environments and has become a hotspot of the robotics research community [4]. Various sensors (e.g., radar, ultrasonic, laser) have been applied in SLAM to help a robot perceive information in an unknown scene [5]. In comparison to the ranging type of sensors, a camera with advantages of smaller size and low power consumption can provide abundant texture information for robots in unknown environments, especially in GNSS-denied (Global Navigation Satellite System) environments (e.g., lunar, Martian surface, and underground environments). Coinciding with the development of computer technology, visual SLAM has attracted extensive attentions and has been applied successfully to planetary exploration missions such as the NASA’s

Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call