TY - GEN
T1 - Depth Completion via A Dual-Fusion Method
AU - Yang, Shuwen
AU - Xiao, Luwei
AU - Zhang, Junhang
AU - Fu, Zhichao
AU - Ma, Tianlong
AU - He, Liang
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Depth completion technology based on multi-level feature fusion (MF) strategy has recently achieved remarkable success. However, the existing MF-based methods treat RGB features and depth map features equally when performing modal fusion but ignore the difference in semantic richness and sparsity between them, which leads to the results generated by these methods overfitting the shape of RGB and harm to the accuracy of depth value. To address this problem, we proposed a novel dual fusion (DF) strategy for MF-based depth completion, which can prevent overfitting by weakening the influence of RGB features on the generated results through two fusion stages. The entire DF framework consists of two multi-level fusion modules. The first fusion module performs a simple fusion of RGB features and depth features, while the second fusion module enriches the sparse image representation with the previously obtained fused features. Besides, we utilize non-local sparse attention to solve the problem that ordinary convolution is not capable of expressing depth map features enough. We test our approach on the outdoor KITTI test set and achieve the state-of-the-art (SOTA) performance in RMSE. Extensive experiments on the indoor NYUv2 dataset and KITTI validation set further demonstrate that our approach outperforms existing MF-based methods.
AB - Depth completion technology based on multi-level feature fusion (MF) strategy has recently achieved remarkable success. However, the existing MF-based methods treat RGB features and depth map features equally when performing modal fusion but ignore the difference in semantic richness and sparsity between them, which leads to the results generated by these methods overfitting the shape of RGB and harm to the accuracy of depth value. To address this problem, we proposed a novel dual fusion (DF) strategy for MF-based depth completion, which can prevent overfitting by weakening the influence of RGB features on the generated results through two fusion stages. The entire DF framework consists of two multi-level fusion modules. The first fusion module performs a simple fusion of RGB features and depth features, while the second fusion module enriches the sparse image representation with the previously obtained fused features. Besides, we utilize non-local sparse attention to solve the problem that ordinary convolution is not capable of expressing depth map features enough. We test our approach on the outdoor KITTI test set and achieve the state-of-the-art (SOTA) performance in RMSE. Extensive experiments on the indoor NYUv2 dataset and KITTI validation set further demonstrate that our approach outperforms existing MF-based methods.
UR - https://www.scopus.com/pages/publications/85143590969
U2 - 10.1109/ICPR56361.2022.9956653
DO - 10.1109/ICPR56361.2022.9956653
M3 - 会议稿件
AN - SCOPUS:85143590969
T3 - Proceedings - International Conference on Pattern Recognition
SP - 3686
EP - 3691
BT - 2022 26th International Conference on Pattern Recognition, ICPR 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 26th International Conference on Pattern Recognition, ICPR 2022
Y2 - 21 August 2022 through 25 August 2022
ER -