2021
DOI: 10.1109/tip.2021.3079821
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive Context-Aware Multi-Modal Network for Depth Completion

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
66
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 125 publications
(84 citation statements)
references
References 64 publications
0
66
0
Order By: Relevance
“…For dense estimation, Xiong et al [12] explored graph construction on 3D points to guide depth completion. Further, Zhao et al [36] directly constructed an encoder-decoder architecture based on graph propagation to explore better multi-modality feature fusion for depth completion. Although graph-based methods generally have fewer parameters, large-scale graph construction and messages aggregation require heavy computations, which limits general applications.…”
Section: Related Workmentioning
confidence: 99%
“…For dense estimation, Xiong et al [12] explored graph construction on 3D points to guide depth completion. Further, Zhao et al [36] directly constructed an encoder-decoder architecture based on graph propagation to explore better multi-modality feature fusion for depth completion. Although graph-based methods generally have fewer parameters, large-scale graph construction and messages aggregation require heavy computations, which limits general applications.…”
Section: Related Workmentioning
confidence: 99%
“…Besides, PI-RCNN fuses image features into point clouds with their new PACF module. Other methods, such as EPNet [27] and ACMNet [28], adopt a gating mechanism to control the portion of LiDAR and image information during fusion.…”
Section: A Multimodal Approaches For 3d Tasksmentioning
confidence: 99%
“…Although these previous works have demonstrated satisfactory results at the time they were published [28], [29], some later single-modal methods have surpassed them [30], [31]. Moreover, such extra performance often comes at the expense of higher computation and memory costs due to the extra inputs, thus incurring a higher hurdle for online tasks than their offline counterparts.…”
Section: A Multimodal Approaches For 3d Tasksmentioning
confidence: 99%
“…Except the application of spatial propagation scheme, Tang et al [19] proposed GuideNet to fuse the LiDAR data and RGB image information by performing learnable content-dependent and spatially-variant kernels. Zhao et al [25] proposed to adopt the graph propagation to model the observed spatial contexts with depth values, so as to better guide the recovery of the unobserved pixels' depth. More recently, PENet [11] and FCFR-Net [13] were proposed to carried out the depth completion through a two-stage coarse-to-fine mechanism and achieved impressive results.…”
Section: Depth Completion Based On Multimodal Learningmentioning
confidence: 99%