z-logo
open-access-imgOpen Access
Fusion of dense spatial features and sparse temporal features for three‐dimensional structure estimation in urban scenes
Author(s) -
Nawaf Mohamad Motasem,
Trémeau Alain
Publication year - 2013
Publication title -
iet computer vision
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.38
H-Index - 37
eISSN - 1751-9640
pISSN - 1751-9632
DOI - 10.1049/iet-cvi.2012.0270
Subject(s) - artificial intelligence , computer vision , computer science , structure from motion , markov random field , optical flow , motion estimation , segmentation , pattern recognition (psychology) , image segmentation , image (mathematics)
The authors present a novel approach to improve three‐dimensional (3D) structure estimation from an image stream in urban scenes. The authors consider a particular setup, where the camera is installed on a moving vehicle. Applying traditional structure from motion (SfM) technique in this case generates poor estimation of the 3D structure because of several reasons such as texture‐less images, small baseline variations and dominant forward camera motion. The authors idea is to introduce the monocular depth cues that exist in a single image, and add time constraints on the estimated 3D structure. The scene is modelled as a set of small planar patches obtained using over‐segmentation, and the goal is to estimate the 3D positioning of these planes. The authors propose a fusion scheme that employs Markov random field model to integrate spatial and temporal depth features. Spatial depth is obtained by learning a set of global and local image features. Temporal depth is obtained via sparse optical flow based SfM approach. That allows decreasing the estimation ambiguity by forcing some constraints on camera motion. Finally, the authors apply a fusion scheme to create unique 3D structure estimation.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here