2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.01101
|View full text |Cite
|
Sign up to set email alerts
|

What You See is What You Get: Exploiting Visibility for 3D Object Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
54
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 124 publications
(60 citation statements)
references
References 16 publications
0
54
0
Order By: Relevance
“…During the measurement process, the detected LiDAR points are the result of a physical raycasting. When representing the LiDAR points using pillar features, one fundamentally neglects the hidden model information of observability, including information on free space and occupied areas [8]. However, we argue that the observability information might be beneficial for the dense top-view segmentation.…”
Section: A Point Cloud Feature Encodingmentioning
confidence: 88%
See 1 more Smart Citation
“…During the measurement process, the detected LiDAR points are the result of a physical raycasting. When representing the LiDAR points using pillar features, one fundamentally neglects the hidden model information of observability, including information on free space and occupied areas [8]. However, we argue that the observability information might be beneficial for the dense top-view segmentation.…”
Section: A Point Cloud Feature Encodingmentioning
confidence: 88%
“…The commonly used 2D occupancy grid map encodes the occupancy probability for each evenly spaced grid cell on the ground plane. Since the mapping of 3D measurements to 2D implies a loss of information in the height, [8] further divides the 3D world into a set of 3D voxels and encodes the occupancy information for each voxel to obtain 3D occupancy grid maps. In addition to occupancy, other features such as intensity, density and observations can also be derived to form multi-layer grid maps [9].…”
Section: A Point Cloud Representationmentioning
confidence: 99%
“…3D Object Detection Modern LiDAR-based 3D object detectors can be organized into three sub-categories based on the way they represent the input point cloud: i.e., voxelization-based detectors [55,8,21,50,19,60,47,68,58,20,64,56], point-based methods [45,63,32,38,62,46] as well as hybrid methods [67,61,5,12,44]. Besides input representation, aggregating points across frames [13,65,14,41], using additional input modalities [19,4,39,57,25,29,48,37], and multi-task training [27,59,30,24] have also been studied to boost the performance. Despite such progress in model design, the output representation and evaluation metrics have remained mostly unchanged.…”
Section: Related Workmentioning
confidence: 99%
“…In our literature search, we find only Huang et al [16] have tackled this issue previously. Ngiam et al [24] and Hu et al [15] also consider multiple 3D frames as input, but both use relatively simple techniques of reusing seed points or concatenating input over multiple frames.…”
Section: Introductionmentioning
confidence: 99%
“…Sequences 0,1,3,4,5,9,11,12,15,17,19,20 were used for training, while the remaining were chosen for validation.…”
mentioning
confidence: 99%