Abstract

Background: Navigation and localization are key to the successful execution of autonomous unmanned underwater vehicles (UUVs) in marine environmental monitoring, underwater 3D mapping, and ocean resource surveys. The estimation of the position and the orientation of autonomous UUVs are a long-standing challenging and fundamental problem. As one of the underwater sensors, camera has always been the focus of attention due to its advantages of low cost and rich content information in low visibility waters, especially in the fields of visual perception of the underwater environment, target recognition and tracking. At present, the visual real-time pose estimation technology that can be used for UUVs is mainly divided into geometry-based visual positioning algorithms and deep learning-based visual positioning algorithms. Methods: In order to compare the performance of different positioning algorithms and strategies, this paper uses C++ and python, takes the ORB-SLAM3 algorithm and DF-VO algorithm as representatives to conduct a comparative experiment and analysis. Results: The geometry-based algorithm ORB-SLAM3 is less affected by illumination, performs more stably in different underwater environments, and has a shorter calculation time, but its robustness is poor in complex environments. The visual positioning algorithm DF-VO based on deep learning takes longer time to compute, and the positioning accuracy is more easily affected by illumination, especially in dark conditions. However, its robustness is better in unstructured environments such as large-scale image rotation and dynamic object interference. Conclusions: In general, the deep learning-based algorithm is more robust, but multiple deep learning networks make it need more time to compute. The geometry-based method costs less time and is more accurate in low-light and turbid underwater conditions. However, in real underwater situations, these two methods can be connected as binocular vision or methods of multi-sensor combined pose estimation.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call