Abstract

In this paper, we present a novel deep learning based method for video anomaly detection and localization. The key idea of our approach is that the latent space representations of normal samples are trained to accord with a specific prior distribution by the proposed deep neural network - Multivariate Gaussian Fully Convolution Adversarial Autoencoder (MGFC-AAE), while the latent representations of anomalies do not. In order to extract deep features from input samples as latent representations, a convolutional neural network (CNN) is employed for the encoder of the deep network. Based on the probability that the test sample is associated with the prior distribution, an energy-based method is applied to obtain its anomaly score. A two-stream framework is utilized to integrate the appearance and motion cues to achieve more comprehensive detection results, taking the gradient and optical flow patches as inputs for each stream. Besides, a multi-scale patch structure is put forward to handle the perspective of some video scenes. Experiments are conducted on three public datasets, results verify that our framework can accurately detect and locate abnormal objects in various video scenes, achieving competitive performance when compared with other state-of-the-art works.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call