Abstract
This paper proposes a vision-based Semantic Unscented FastSLAM (UFastSLAM) algorithm for mobile robot combing the semantic relationship and the unscented FastSLAM. The landmarks are detected by a binocular vision, and the semantic observation model can be created by transforming the semantic relationships into the semantic metric map. Semantic Unscented FastSLAM can be used to update the localization of the landmarks and robot pose even when the encoders inherits large accummative errors that may not be corrected by the loop closure detection of the vision system Experiments have been carried out to demonstrate that the Semantic Unscented FastSLAM algorithm can achieve much better performance in indoor autonomous survalience than Unscented FastSLAM.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.