Abstract

Vision based mapping is an emerging technology with decades of research advancements. The most famous mapping method available is silmuntaneous localization and mapping (SLAM) which provide an accurate map projected in a simulation. Unfortunately, SLAM requires an active sensor in order to acquire the data from its environment opposite the vision-based mapping which requires a passive sensor to collect data. This project aims to develop an autonomous mapping and exploration algorithm, design a controller for the robot-tracked vehicle and analyze the accuracy of the algorithm. The problem in autonomous mapping is precision, limitation of computational power and complex computation. So, the algorithm will be based on the visual odometer algorithm through a single-visual sensor. The robot tracker has also been designed and implemented on Raspberry Pi 3. The accuracy of two object with different height was calculated to ensure the validation of the algorithm being able to project the real object in 3D projection. The result for the task is shown in figures as to present the capability of the algorithm in projecting the map in 3D projection. The algorithm works as expected but still requires improvements to increase the precision of the map projection.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call