Abstract

SLAM in dynamic environments is still a severe challenge for most feature-based SLAM systems. Moving objects will lead to terrible errors in the calculation of frame tracking and local mapping. We propose a novel method for keypoints selection to lower the negative effect brought by moving objects during map construction. To address this challenge, we concentrate on the combination of coarse semantic information and a feature-based SLAM system. In this article, a modified CenterNet object detector is proposed as the moving object detection thread for providing coarse semantic information and 2D location. The modified CenterNet can provide a faster and more accurate prediction. For each frame in a sequence, objects in a scenario will be classified into two motion states, non-static and static, according to the category prediction from the moving object detection thread. Then important processing called semantic data association is presented for motion removal in frame tracking thread and local mapping thread. In this way, the keypoints on the non-static objects will not be chosen for calculation. Finally, on this basis, a modified real-time SLAM system is presented. Experimental results in TUM indoor dataset show that the proposed SLAM system outperforms the state-of-the-art ORB-SLAM2 in dynamic environments. Besides that, our method has a better speed-accuracy trade-off than other deep learning-based SLAM systems in dynamic environments.

Highlights

  • In recent years, Simultaneous Localization and Mapping (SLAM) has been a hot research topic in the Computer Vision and Robotics fields and has recently been applied to fully autonomous vehicles, including auto-navigation robots, self-driving cars or unmanned aerial vehicles (UAVs) [1], [2]

  • Our proposed SLAM system is built on ORB-SLAM2 which is used as the basic SLAM system

  • The moving object detection thread is based on our modified CenterNet

Read more

Summary

Introduction

Simultaneous Localization and Mapping (SLAM) has been a hot research topic in the Computer Vision and Robotics fields and has recently been applied to fully autonomous vehicles, including auto-navigation robots, self-driving cars or unmanned aerial vehicles (UAVs) [1], [2]. SLAM is an assembly of many techniques to achieve the goal that moving robots equipped with special sensors can build the surrounding scenes and evaluate movement poses without prior environment information [3]. In the field of robot auto-navigation and autonomous driving, a prebuilt. The early SLAM research focuses on the Lidar sensor because of its measuring accuracy and easier data process. Compared with the Lidar sensor, the visual sensor can provide more representation information, perform better in complex scenarios and have better performance in loop detection

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call