Abstract

Cameras are one of the most relevant sensors in autonomous robots. However, two of their challenges are to extract useful information from captured images, and to manage the small field of view of regular cameras. This paper proposes implementing a dynamic visual memory to store the information gathered from a moving camera on board a robot, followed by an attention system to choose where to look with this mobile camera, and a visual localization algorithm that incorporates this visual memory. The visual memory is a collection of relevant task-oriented objects and 3D segments, and its scope is wider than the current camera field of view. The attention module takes into account the need to reobserve objects in the visual memory and the need to explore new areas. The visual memory is useful also in localization tasks, as it provides more information about robot surroundings than the current instantaneous image. This visual system is intended as underlying technology for service robot applications in real people's homes. Several experiments have been carried out, both with simulated and real Pioneer and Nao robots, to validate the system and each of its components in office scenarios.

Highlights

  • Computer vision research is growing rapidly, both in robotics and in many other applications, from surveillance systems for security to the automatic acquisition of 3D models for Virtual Reality displays.The number of commercial applications is increasing, such as traffic monitoring, parking entrance control, augmented reality video games, and face recognition

  • Regular cameras typically have 60 degrees of scope. This would be good enough for visual control but a broader scope may improve robot responses in tasks like navigation, where the presence of obstacles in the robot’s surroundings should be taken into account even if they lie outside the current field of view

  • If the uncertainty on a 3D segment falls below a given threshold, it is deleted from visual memory

Read more

Summary

Introduction

Computer vision research is growing rapidly, both in robotics and in many other applications, from surveillance systems for security to the automatic acquisition of 3D models for Virtual Reality displays. The control algorithm should be robust enough to face the lack of time persistence of relevant stimuli in images This poses a challenge when the objects lie beyond the current field of view of the camera. The visual representation of interesting objects around the robot beyond the current field of view may improve the quality of the robot’s behavior as it handles more information when making decisions This is where the problem of selecting where to look every time, known as gaze control, or overt attention [2,3] arises. A visual localization algorithm has been developed which uses the current image or the contents of the memory to continuously estimate the robot position It provides a robust localization estimation and has been designed to handle symmetries in the environment.

Related Works
Design
Local Visual Memory
Reconstruction with 3D Segments
Inserting Segments into the 3D Visual Memory
Complex Primitives in Visual Memory
Visual Attention
Gaze Control
Tracking of a Focused Object
Exploring New Areas of Interest
Representation of the Environment
Attention Module Operation
Evolutionary Visual Localization
Analyzing Images
Health Calculation from Instantaneous Images
Health Calculation with Visual Memory
Explorer Creation
Race Maanagement
Race Evolution
Selecting the Robot Pose Estimation
Experiments
Robot in the Middle of a Room
Robot Navigating a Curve
Robot Occlusions
Attentive Visual Memory on a Humanoid Robot
Visual Localization Experiments
Testing MCL Algorithm Behavior
Typical Execution in Humanoid Robot
Dealing with Symmetries and Kidnappings
Health Function Based on Instantaneous Images
Health Function Based on Visual Memory
Conclusions
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.