2020
DOI: 10.3390/rs12111895
|View full text |Cite
|
Sign up to set email alerts
|

KDA3D: Key-Point Densification and Multi-Attention Guidance for 3D Object Detection

Abstract: In this paper, we propose a novel 3D object detector KDA3D, which achieves high-precision and robust classification, segmentation, and localization with the help of key-point densification and multi-attention guidance. The proposed end-to-end neural network architecture takes LIDAR point clouds as the main inputs that can be optionally complemented by RGB images. It consists of three parts: part-1 segments 3D foreground points and generates reliable proposals; part-2 (optional) enhances point cloud density and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 20 publications
(7 citation statements)
references
References 38 publications
0
6
0
Order By: Relevance
“…The results for car validation and test set are given in Tables 21 and 22 respectively while for pedestrian and cyclist categories on validation set are listed in Table 23. Wen et al [77] illustrates that the proposed model [77] competes with [199,201] in comprehensive performance. For the cyclist class, it outperforms the [201] while in the car class, it is 2× faster than [201].…”
Section: Performance Evaluationmentioning
confidence: 98%
See 1 more Smart Citation
“…The results for car validation and test set are given in Tables 21 and 22 respectively while for pedestrian and cyclist categories on validation set are listed in Table 23. Wen et al [77] illustrates that the proposed model [77] competes with [199,201] in comprehensive performance. For the cyclist class, it outperforms the [201] while in the car class, it is 2× faster than [201].…”
Section: Performance Evaluationmentioning
confidence: 98%
“…Wen et al [77] illustrates that the proposed model [77] competes with [199,201] in comprehensive performance. For the cyclist class, it outperforms the [201] while in the car class, it is 2× faster than [201].…”
Section: Performance Evaluationmentioning
confidence: 98%
“…As for the LiDAR branch mentioned above, the point cloud can be used in the form of 3D points with reflectance, voxelized tensor, front-view/ range-view/ bird's eye view, as well as pseudo-point clouds. Though all these data have different intrinsic characteristics, which are highly associated with the latter LiDAR backbone, most of these data are produced with a rule-based procession except pseudo-point clouds [79]. Besides, all these data representations of Li-DAR can be visualized straightforward because the data in this stage still have interpretability compared to the embedding in feature space.…”
Section: Strong-fusionmentioning
confidence: 99%
“…[54] transforms 3D lidar point clouds into the 2D image and fuses feature-level representation in the image branch leveraging mature CNN techniques to achieve better performance. [87] fused raw RGB pixel with the voxelized tensor while [79] directly combines pseudo-point clouds generated from the image branch and raw point clouds from the LiDAR branch together to accomplish object detection tasks.…”
Section: Images Semantic Segmentationmentioning
confidence: 99%
“…Since the semantic understanding and analysis of a 3D point cloud is the basis for realizing scene understanding [1,2], the application of semantic segmentation of 3D point cloud has been more and more extensive in recent years [3][4][5], such as augmented/virtual reality [6] and intelligent robot [7]. Moreover, in the field of self-driving, the accurate perception of the environment based on LIDAR point cloud data is the key to realize information decision-making and driving safely in the complex dynamic environment.…”
Section: Introductionmentioning
confidence: 99%