Abstract
We present a framework to perform novelty detection using visual input in which a mobile robot first learns a model of normality in its operating environment and later uses this to highlight uncommon visual features that may appear. This ability is of great importance for both robotic exploration and inspection tasks, because it enables the robot to allocate computational and attentional resources efficiently to those features which are novel. At the heart of the proposed system is the image encoding mechanism which uses local colour statistics from regions selected by a biologically-inspired model of visual attention. Our approach works in real-time with a wide, unrestricted field of view and is robust to image transformations. Experiments conducted in an engineered scenario demonstrate the efficiency and functionality of our method.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.