z-logo
open-access-imgOpen Access
Monocular Depth Estimation via Convolutional Neural Network with Attention Module
Author(s) -
Lawrence W. Lan,
Yaping Zhang,
Yuwei Yang
Publication year - 2021
Publication title -
journal of physics. conference series
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.21
H-Index - 85
eISSN - 1742-6596
pISSN - 1742-6588
DOI - 10.1088/1742-6596/2025/1/012062
Subject(s) - computer science , feature (linguistics) , convolutional neural network , artificial intelligence , monocular , depth map , channel (broadcasting) , rgb color model , encoder , computer vision , pattern recognition (psychology) , feature vector , field (mathematics) , image (mathematics) , mathematics , computer network , philosophy , linguistics , pure mathematics , operating system
The depth accurately estimated from the RGB image can be used in applications such as 3D reconstruction and scene display. In depth estimation, Convolutional Neural Network (CNN) plays an important role. However, most of the existing CNN-based depth estimation methods often do not take full advantage of the space and channel information in the local receptive field, resulting in lower-resolution depth maps. Based on the extant Encoder-Decoder network architecture, a simple and effective attention module is introduced to learn the discerning feature and the semantic feature in this paper. First, the intermediate feature map is generated by the encoder, and then the attention module is used to generate the attention maps along the two dimensions of the space and channel of the feature map. Finally, the refined output generated by multiplying the attention maps to intermediate feature maps is input to the subsequent neural network for further work. Our works show that the suggest approach has obtained good achievements on NYU Depth v2.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here