Abstract

Abstract. Background subtraction aims at detecting salient background which in return provides regions of moving objects referred to as the foreground. Background subtraction inherently uses the temporal relations by including time dimension in its formulation. Alternative techniques to background subtraction require stationary cameras for learning the background. Stationary cameras provide semi-constant background images that make learning salient background easier. Still cameras, however, are not applicable to moving camera scenarios, such as vehicle embedded camera for autonomous driving. For moving cameras, due to the complexity of modelling changing background, recent approaches focus on directly detecting the foreground objects in each frame independently. This treatment, however, requires learning all possible objects that can appear in the field of view. In this paper, we achieve background subtraction for moving cameras using specialized deep learning approach, the Moving-camera Background Subtraction Network (MBS-Net). Our approach is robust to detect changing background in various scenarios and does not require training on foreground objects. The developed approach uses temporal cues from past frames by applying Conditional Random Fields as a part of the developed neural network. Our proposed method have a good performance on ApolloScape dataset (Huang et al., 2018) with resolution 3384 × 2710 videos. To the best of our acknowledge, this paper is the first to propose background subtraction for moving cameras using deep learning.

Highlights

  • Primary goal of background subtraction is to find moving objects based on their differences from the salient background which is learned from a stream of images

  • We model the background using convolutional neural networks (CNN) that has been successfully applied to image segmentation among others to model complex and recessive relationships between the inputs and outputs (Vemulapalli et al, 2016)

  • We introduce the Conditional Random Fields (CRFs) as temporal regularization to ensure the background information learned in the previous frames is carried over to the current CNN output

Read more

Summary

Introduction

Primary goal of background subtraction is to find moving objects based on their differences from the salient background which is learned from a stream of images. This task can be considered as classification of each pixel as background or foreground that can be designated as a pixel-wise binary semantic segmentation task. Semantic background subtraction, has shown to improve the performance of background subtraction These methods aim to label pixels into a number of moving object regions (Cioppa et al, 2020, Braham et al, 2017). We model the background using convolutional neural networks (CNN) that has been successfully applied to image segmentation among others to model complex and recessive relationships between the inputs and outputs (Vemulapalli et al, 2016)

Objectives
Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.