2020
DOI: 10.1109/tip.2019.2960589
|View full text |Cite
|
Sign up to set email alerts
|

HMS-Net: Hierarchical Multi-Scale Sparsity-Invariant Network for Sparse Depth Completion

Abstract: Dense depth cues are important and have wide applications in various computer vision tasks. In autonomous driving, LIDAR sensors are adopted to acquire depth measurements around the vehicle to perceive the surrounding environments. However, depth maps obtained by LIDAR are generally sparse because of its hardware limitation. The task of depth completion attracts increasing attention, which aims at generating a dense depth map from an input sparse depth map. To effectively utilize multi-scale features, we propo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
89
0
2

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 119 publications
(98 citation statements)
references
References 42 publications
1
89
0
2
Order By: Relevance
“…Although we mainly focus on the outdoor application scenarios, we also train our model on indoor scenes, i.e., NYU-Depth-v2 dataset. As NYU-Depth-v2 dataset provides relatively denser depth measurements by Microsoft Kinect, we uniformly sample the depth map to obtain the sparser version following previous works [26,14]. We compare our results with latest CNN-based methods [26,2] as well as the classic methods [21,35,20] as shown in Table 3, and our method achieves state-of-the-art performance as well.…”
Section: Analysis Of Generalization Ability and Stabilitymentioning
confidence: 95%
See 1 more Smart Citation
“…Although we mainly focus on the outdoor application scenarios, we also train our model on indoor scenes, i.e., NYU-Depth-v2 dataset. As NYU-Depth-v2 dataset provides relatively denser depth measurements by Microsoft Kinect, we uniformly sample the depth map to obtain the sparser version following previous works [26,14]. We compare our results with latest CNN-based methods [26,2] as well as the classic methods [21,35,20] as shown in Table 3, and our method achieves state-of-the-art performance as well.…”
Section: Analysis Of Generalization Ability and Stabilitymentioning
confidence: 95%
“…Ma et al concatenated the sparse depth and color image as the inputs of an off-the-shelf network [26] and further explored the feasibility of self-supervised Li-DAR completion [23]. Moreover, [14,16,33,4] proposed different network architectures to better exploit the potential of the encoder-decoder framework. However, the encoderdecoder architecture tends to predict the depth maps comprehensively but fails to concentrate on the local areas.…”
Section: Related Workmentioning
confidence: 99%
“…Supervised Depth Completion minimizes the discrepancy between ground truth depth and depth predicted from an RGB image and sparse depth measurements. Methods focus on network topology [14,25,28], optimization [3,4,30], and modeling [5,8]. To handle sparse depth, [14] employed early fusion, where the image and sparse depth are convolved separately and the results concatenated as the input to a ResNet encoder.…”
Section: Related Workmentioning
confidence: 99%
“…[5] proposed a normalized convolutional layer to propagate sparse depth and used a binary validity map as a confidence measure. [8] proposed an upsampling layer and joint concatenation and convolution to deal with sparse inputs. All these methods require per-pixel ground-truth annotation.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation