z-logo
open-access-imgOpen Access
Sparse Deep Feature Representation for Object Detection from Wearable Cameras
Author(s) -
Quanfu Fan,
Richard J. Chen
Publication year - 2017
Language(s) - English
Resource type - Conference proceedings
DOI - 10.5244/c.31.163
Subject(s) - computer science , artificial intelligence , wearable computer , representation (politics) , computer vision , feature (linguistics) , object (grammar) , object detection , feature extraction , sparse approximation , wearable technology , pattern recognition (psychology) , embedded system , linguistics , philosophy , politics , political science , law
We propose a novel sparse feature representation for the faster RCNN framework and apply it for object detection from wearable cameras. Two main ideas, sparse convolution and sparse ROI pooling, are developed to reduce model complexity as well as computational cost. Sparse convolution approximates a full kernel by skipping weights in the kernel while sparse ROI pooling performs feature dimensionality reduction on the ROI pooling layer by skipping odd-indexed or even-indexed features. We demonstrate the effectiveness of our approach on two challenging body camera datasets including realistic police-generated clips. Our approach achieves a significant reduction of model size by a factor of over 10× as well as a computational speedup of about 2×, yet without compromising much detection accuracy compared to a VGG16-based baseline detector.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom