2020
DOI: 10.1109/access.2020.2990212
|View full text |Cite
|
Sign up to set email alerts
|

Deep Architecture With Cross Guidance Between Single Image and Sparse LiDAR Data for Depth Completion

Abstract: It is challenging to apply depth maps generated from sparse laser scan data to computer vision tasks, such as robot vision and autonomous driving, because of the sparsity and noise in the data. To overcome this problem, depth completion tasks have been proposed to produce a dense depth map from sparse LiDAR data and a single RGB image. In this study, we developed a deep convolutional architecture with cross guidance for multi-modal feature fusion to compensate for the lack of representation power of their moda… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
17
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 39 publications
(19 citation statements)
references
References 33 publications
(69 reference statements)
0
17
0
Order By: Relevance
“…CFCNet [8] learns to capture the semantically correlated features between RGB and depth information. CG-Net [9] proposes a cross guidance module to fuse the multi-modal feature from RGB and LiDAR. Some methods rely on iterative Spatial Propagation Network (SPN) to better treat the difficulties made by the sparsity and irregular distribution of the input [10][11][12] [13].…”
Section: Related Workmentioning
confidence: 99%
“…CFCNet [8] learns to capture the semantically correlated features between RGB and depth information. CG-Net [9] proposes a cross guidance module to fuse the multi-modal feature from RGB and LiDAR. Some methods rely on iterative Spatial Propagation Network (SPN) to better treat the difficulties made by the sparsity and irregular distribution of the input [10][11][12] [13].…”
Section: Related Workmentioning
confidence: 99%
“…Some methods, e.g. [45], fuse the sparse depth and RGB image via early fusion while others [18,26,29,37,44,64] utilize a late fusion scheme, or jointly utilize both the early and late fusion [36,65,68]. Another line of research focuses on utilizing affinity or geometric information of the scene via surface normal, occlusion boundaries, and the geometric convolutional layer [11,12,25,34,52,54,77,86].…”
Section: Related Workmentioning
confidence: 99%
“…These solutions train their normal prediction with synthetic data [10] or based on principal component analysis [11]. Others use confidence to fuse image and depth features, giving more weight to the modality with less uncertainty [12], [13], [14].…”
Section: A State Of the Art -Depth Completionmentioning
confidence: 99%
“…Various networks integrate Spatial Pyramid Pooling (SPP) [22] for depth completion [5], [16], [18], [23]. Atrous Spatial Pyramid Pooling (ASPP) [24] has been studied at the end of an encoder [16] or within residual blocks [13]. Li et al [25] combine multiple networks, each using different resolutions of sparse input.…”
Section: A State Of the Art -Depth Completionmentioning
confidence: 99%