Learning Deep Representations of Appearance and Motion for Anomalous Event Detection
Author(s) -
Dan Xu,
Elisa Ricci,
Yan Yan,
Jingkuan Song,
Nicu Sebe
Publication year - 2015
Language(s) - English
Resource type - Conference proceedings
DOI - 10.5244/c.29.8
Subject(s) - artificial intelligence , computer science , motion (physics) , anomaly detection , exploit , event (particle physics) , representation (politics) , pattern recognition (psychology) , feature learning , deep learning , feature (linguistics) , fusion , computer vision , physics , quantum mechanics , linguistics , philosophy , computer security , politics , political science , law
We present a novel unsupervised deep learning framework for anomalous event detection in complex video scenes. While most existing works merely use hand-crafted appearance and motion features, we propose Appearance and Motion DeepNet (AMDN) which utilizes deep neural networks to automatically learn feature representations. To exploit the complementary information of both appearance and motion patterns, we introduce a novel double fusion framework, combining both the benefits of traditional early fusion and late fusion strategies. Specifically, stacked denoising autoencoders are proposed to separately learn both appearance and motion features as well as a joint representation (early fusion). Based on the learned representations, multiple one-class SVM models are used to predict the anomaly scores of each input, which are then integrated with a late fusion strategy for final anomaly detection. We evaluate the proposed method on two publicly available video surveillance datasets, showing competitive performance with respect to state of the art approaches.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom