
An aggregated deep convolutional recurrent model for event based surveillance video summarisation: A supervised approach
Author(s) -
U. Sreeja M.,
Kovoor Binsu C.
Publication year - 2021
Publication title -
iet computer vision
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.38
H-Index - 37
eISSN - 1751-9640
pISSN - 1751-9632
DOI - 10.1049/cvi2.12044
Subject(s) - computer science , convolutional neural network , artificial intelligence , feature extraction , pattern recognition (psychology) , frame (networking) , deep learning , event (particle physics) , machine learning , physics , quantum mechanics , telecommunications
Surveillance video summarisation is characterised by extracting video segments containing abnormal events from surveillance video footages. Accurate identification of abnormal events from surveillance footages is of paramount importance in surveillance video summarisation. Accordingly, the proposed framework builds an aggregated convolutional recurrent model that can precisely detect the suspicious events in a surveillance footage, by employing a supervised learning which is found to yield better results compared with unsupervised counterparts. The preliminary stage in the model is a multilayer Convolutional Neural Network for frame‐level feature extraction followed by stacked bidirectional Gated Recurrent Unit for sequence‐level feature extraction and classification. Since the video clips used for training are not implicit to surveillance, a block‐based approach for testing on surveillance videos is proposed. The results evaluated on two custom datasets, Streets and Campus, prove that the proposed model produces remarkable results leveraging the properties of bidirectional GRU with supervised learning. Extensive experimental analysis on selection of optimum architecture is conducted which substantiates the significance of stacked bidirectional GRUs over unidirectional ones. Additionally, qualitative results ensure that summaries produced are concise, representative, complete, diverse and informative. Moreover, comparison of the performance of the proposed model with state of the art certainly proves the superiority of the proposed model.