
Camera matching based on spatiotemporal activity and conditional random field model
Author(s) -
Xiaokai Liu,
Hongyu Wang,
Hongbo Gao
Publication year - 2014
Publication title -
iet computer vision
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.38
H-Index - 37
eISSN - 1751-9640
pISSN - 1751-9632
DOI - 10.1049/iet-cvi.2013.0160
Subject(s) - artificial intelligence , scale invariant feature transform , computer science , conditional random field , matching (statistics) , computer vision , pixel , pattern recognition (psychology) , a priori and a posteriori , feature extraction , feature (linguistics) , mathematics , statistics , philosophy , epistemology , linguistics
In this study, the authors investigate how to find correspondences (pixel‐level matches) between two video sequences with overlapping views, recorded by different stationary uncalibrated video cameras, without imposing any constrains on, or requiring any a priori knowledge of, the viewing or the illumination conditions under which these videos were obtained. They propose a matching method using a low‐level feature, spatiotemporal activity (STA), which effectively combines both the photometric features and the statistic features and makes the proposed method robust to the pose, illumination and geometric effects. They formulate the author's matching problem as a conditional random field model and optimise it by using graph cuts, instead of applying exhaustive pixel‐wise matching. A qualitative analysis of the performance improvement compared with the temporal activity‐based matching method (TAM) is given by inference from a mathematical model and the quantitative results compared with the TAM and the scale‐invariant feature transform (SIFT) are presented with real life examples. The experiments show that their method significantly outperforms the state‐of‐the‐art methods with a higher accuracy and is orders of magnitudes faster than the comparable ones.