Adapting Object Detectors from Images to Weakly Labeled Videos
Author(s) -
Omit Chanda,
Eu Wern Teh,
Mrigank Rochan,
Zhenyu Guo,
Yang Wang
Publication year - 2017
Language(s) - English
Resource type - Conference proceedings
DOI - 10.5244/c.31.56
Subject(s) - computer vision , computer science , artificial intelligence , detector , object (grammar) , computer graphics (images) , telecommunications
Due to the domain shift between images and videos, standard object detectors trained on images usually do not perform well on videos. At the same time, it is difficult to directly train object detectors from video data due to the lack of labeled video datasets. In this paper, we consider the problem of localizing objects in weakly labeled videos. A video is weakly labeled if we know the presence/absence of an object in a video (or each frame), but we do not know the exact spatial location. In addition to weakly labeled videos, we assume access to a set of fully labeled images. We incorporate domain adaptation in our framework and adapt the information from the labeled images (source domain) to the weakly labeled videos (target domain). Our experimental results on standard benchmark datasets demonstrate the effectiveness of our proposed approach. Our work can be used for collecting large-scale video datasets for object detection.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom