TY - GEN
T1 - Saliency guided depth prediction from a single image
AU - Wang, Yu
AU - Ma, Lizhuang
N1 - Publisher Copyright:
© 2019 by SCITEPRESS - Science and Technology Publications, Lda.
PY - 2019
Y1 - 2019
N2 - With the recent surge of deep neural networks, depth prediction from a single image has seen substantial progress. Deep regression networks are typically learned from large data without much constraints about the scene structure, thus often leading to uncertainties at discontinuous regions. In this paper, we propose a structure-aware depth prediction method based on two observations: depth is relatively smooth within the same objects, and it is usually easier to model relative depth than model the absolute depth from scratch. Our network first predicts an initial depth map and takes an object saliency map as input, which helps to teach the network to learn depth refinement. Specifically, a stable anchor depth is first estimated from the detected salient objects, and the learning objective is to penalize the difference in relative depth versus the estimated anchor. We show such saliency-guided relative depth constraint unveils helpful scene structures, leading to significant gains on the RGB-D saliency dataset NLPR and depth prediction dataset NYU V2. Furthermore, our method is appealing in that it is pluggable to any depth network and is trained end-to-end with no overhead of time during testing.
AB - With the recent surge of deep neural networks, depth prediction from a single image has seen substantial progress. Deep regression networks are typically learned from large data without much constraints about the scene structure, thus often leading to uncertainties at discontinuous regions. In this paper, we propose a structure-aware depth prediction method based on two observations: depth is relatively smooth within the same objects, and it is usually easier to model relative depth than model the absolute depth from scratch. Our network first predicts an initial depth map and takes an object saliency map as input, which helps to teach the network to learn depth refinement. Specifically, a stable anchor depth is first estimated from the detected salient objects, and the learning objective is to penalize the difference in relative depth versus the estimated anchor. We show such saliency-guided relative depth constraint unveils helpful scene structures, leading to significant gains on the RGB-D saliency dataset NLPR and depth prediction dataset NYU V2. Furthermore, our method is appealing in that it is pluggable to any depth network and is trained end-to-end with no overhead of time during testing.
KW - CNN
KW - Single-image depth prediction
UR - https://www.scopus.com/pages/publications/85107113420
U2 - 10.5220/0008099101530159
DO - 10.5220/0008099101530159
M3 - 会议稿件
AN - SCOPUS:85107113420
T3 - Proceedings of the International Conference on Advances in Computer Technology, Information Science and Communications, CTISC 2019
SP - 153
EP - 159
BT - Proceedings of the International Conference on Advances in Computer Technology, Information Science and Communications, CTISC 2019
A2 - Horng, Wen-Bing
A2 - Yue, Yong
PB - SciTePress
T2 - 2019 International Conference on Advances in Computer Technology, Information Science and Communications, CTISC 2019
Y2 - 15 March 2019 through 17 March 2019
ER -