Abstract

Vision is one of the most important of the senses, and humans use it extensively during navigation. We evaluated different types of image and video frame descriptors that could be used to determine distinctive visual landmarks for localizing a person based on what is seen by a camera that they carry. To do this, we created a database containing over 3 km of video-sequences with ground-truth in the form of distance travelled along different corridors. Using this database, the accuracy of localization—both in terms of knowing which route a user is on—and in terms of position along a certain route, can be evaluated. For each type of descriptor, we also tested different techniques to encode visual structure and to search between journeys to estimate a user’s position. The techniques include single-frame descriptors, those using sequences of frames, and both color and achromatic descriptors. We found that single-frame indexing worked better within this particular dataset. This might be because the motion of the person holding the camera makes the video too dependent on individual steps and motions of one particular journey. Our results suggest that appearance-based information could be an additional source of navigational data indoors, augmenting that provided by, say, radio signal strength indicators (RSSIs). Such visual information could be collected by crowdsourcing low-resolution video feeds, allowing journeys made by different users to be associated with each other, and location to be inferred without requiring explicit mapping. This offers a complementary approach to methods based on simultaneous localization and mapping (SLAM) algorithms.

Highlights

  • Satellite-based global positioning systems (GPS) have been available to consumers for many years

  • Interest has been gathering for the development of systems for indoor position sensing: we might consider this to be the challenge in localization systems (Quigley and Stavens, 2010; Huitl et al, 2012; Shen et al, 2013; Kadous and Peterson, 2013)

  • It is very likely that the size of the errors we have observed can be reduced by incorporating simple motion models and a tracker, in SF GABOR SF GABOR ST GAUSS ST GAUSS ST GABOR ST GABOR

Read more

Summary

Introduction

Satellite-based global positioning systems (GPS) have been available to consumers for many years. Vision-based navigation systems are under active development This might be because such systems do not require special markers to be embedded within the environment. Another reason could be that vision provides an immensely rich source of data, from which estimating position is possible. In emerging applications (Moller et al, 2012) such as mobile domestic robots, merely knowing where a robot is is not enough: a robot often needs some information about its immediate environment in order to take appropriate decisions This includes navigating around obstacles and identifying important objects (e.g. pets)

Methods
Results
Discussion
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.