2018
DOI: 10.48550/arxiv.1808.08685
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

HMS-Net: Hierarchical Multi-scale Sparsity-invariant Network for Sparse Depth Completion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
18
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(19 citation statements)
references
References 40 publications
1
18
0
Order By: Relevance
“…Although we mainly focus on the outdoor application scenarios, we also train our model on indoor scenes, i.e., NYU-Depth-v2 dataset. As NYU-Depth-v2 dataset provides relatively denser depth measurements by Microsoft Kinect, we uniformly sample the depth map to obtain the sparser version following previous works [26,14]. We compare our results with latest CNN-based methods [26,2] as well as the classic methods [21,35,20] as shown in Table 3, and our method achieves state-of-the-art performance as well.…”
Section: Analysis Of Generalization Ability and Stabilitymentioning
confidence: 95%
See 1 more Smart Citation
“…Although we mainly focus on the outdoor application scenarios, we also train our model on indoor scenes, i.e., NYU-Depth-v2 dataset. As NYU-Depth-v2 dataset provides relatively denser depth measurements by Microsoft Kinect, we uniformly sample the depth map to obtain the sparser version following previous works [26,14]. We compare our results with latest CNN-based methods [26,2] as well as the classic methods [21,35,20] as shown in Table 3, and our method achieves state-of-the-art performance as well.…”
Section: Analysis Of Generalization Ability and Stabilitymentioning
confidence: 95%
“…Ma et al concatenated the sparse depth and color image as the inputs of an off-the-shelf network [26] and further explored the feasibility of self-supervised Li-DAR completion [23]. Moreover, [14,16,33,4] proposed different network architectures to better exploit the potential of the encoder-decoder framework. However, the encoderdecoder architecture tends to predict the depth maps comprehensively but fails to concentrate on the local areas.…”
Section: Related Workmentioning
confidence: 99%
“…Instead of simulating a LiDAR via ray-casting, which is computationally expensive and hard to implement [47], we leverage the z-buffer of our synthetic rendering engine to provide a dense depth ground truth at first. Previous approaches used synthetic sparse data to evaluate a model in indoor scenes or synthetic outdoor scenes [6,35,48]. To sparsify the data a Bernoulli distribution per pixel is used in some works [2,35,48] which, given a probability p B and a dense depth image x D , samples each of the pixels x D,k by either keeping the value x D,k with probability p B or setting its value to 0 with probability (1 − p B ), thus generating the sparse depth x s B D .…”
Section: Data Generation Via Projectionsmentioning
confidence: 99%
“…Previous approaches used synthetic sparse data to evaluate a model in indoor scenes or synthetic outdoor scenes [6,35,48]. To sparsify the data a Bernoulli distribution per pixel is used in some works [2,35,48] which, given a probability p B and a dense depth image x D , samples each of the pixels x D,k by either keeping the value x D,k with probability p B or setting its value to 0 with probability (1 − p B ), thus generating the sparse depth x s B D . We argue that using x s B D does not simulate well a real LiDAR input, thus a model trained with x s B D does not perform well in the real domain.…”
Section: Data Generation Via Projectionsmentioning
confidence: 99%
“…Konno et al [39]used residual interpolation to combine low-resolution depth maps with high-resolution RGB images to generate high-resolution depth maps. Based on [9], Huang et al [40]utilized the sparsity-invariant layer to design a sparsity-invariant multi-scale encoder-decoder network for sparse depth completion with RGB guidance. Zhang et al [41]proposed to predict surface normal and occlusion boundary from a deep network and further utilize them to help depth completion in indoor scenes.…”
Section: Guided Depth Upsamplingmentioning
confidence: 99%