2022
DOI: 10.1145/3462219
|View full text |Cite
|
Sign up to set email alerts
|

Multi-feature Fusion VoteNet for 3D Object Detection

Abstract: In this article, we propose a Multi-feature Fusion VoteNet (MFFVoteNet) framework for improving the 3D object detection performance in cluttered and heavily occluded scenes. Our method takes the point cloud and the synchronized RGB image as inputs to provide object detection results in 3D space. Our detection architecture is built on VoteNet with three key designs. First, we augment the VoteNet input with point color information to enhance the difference of various instances in a scene. Next, we integrate an i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 18 publications
(8 citation statements)
references
References 50 publications
0
8
0
Order By: Relevance
“…Most importantly, the advantages of the proposed n-sigmoid CSA-VoteNet stems from two significant facts: on the one hand, it assigns different relevance weights to different elements of the input data during the voting process; on the other hand, the attention map enables the network to emphasize important features and suppress irrelevant or noisy ones. COG [34] 2D-driven [35] F-PointNet [36] VoteNet [12] MLCVNet [37] DeMF [38] CSA-VoteNet A-SCN [39] Point-attention [40] CAA [31] Point-transformer [32] Offset-attention [41] CSA A-SCN [39] Point-attention [40] CAA [31] Pointtransformer [32] Offset-attention [41] CSA…”
Section: Methods In Comparisonmentioning
confidence: 99%
“…Most importantly, the advantages of the proposed n-sigmoid CSA-VoteNet stems from two significant facts: on the one hand, it assigns different relevance weights to different elements of the input data during the voting process; on the other hand, the attention map enables the network to emphasize important features and suppress irrelevant or noisy ones. COG [34] 2D-driven [35] F-PointNet [36] VoteNet [12] MLCVNet [37] DeMF [38] CSA-VoteNet A-SCN [39] Point-attention [40] CAA [31] Point-transformer [32] Offset-attention [41] CSA A-SCN [39] Point-attention [40] CAA [31] Pointtransformer [32] Offset-attention [41] CSA…”
Section: Methods In Comparisonmentioning
confidence: 99%
“…The current mainstream solution for autonomous driving is to use Lidar and vision cameras for fusion [3], Lidar (Light Detection And Ranging) which is short for Lidar. The main Lidar ranging principles are: time-offlight (TOF) method [4], laser triangulation [5], and continuous wave frequency modulation (FMCW) [6], which is a noncontact active ranging sensor [7] [8]. Vision cameras can be further divided into monocular cameras, binocular cameras [39], and depth cameras [10].…”
Section: Introductionmentioning
confidence: 99%
“…The primary goal of three-dimensional (3D) object detection in autonomous driving scenarios is to locate and identify items in road scenes. This task has drawn significant attention from industries 1 , 2 and academia 3 , 4 . Lidar is frequently utilized in the field of autonomous driving and is capable of precisely sensing 3D spatial information.…”
Section: Introductionmentioning
confidence: 99%
“…This task has drawn significant attention from industries 1,2 and academia. 3,4 Lidar is frequently utilized in the field of autonomous driving and is capable of precisely sensing 3D spatial information. Lidar-based 3D object detection has attracted more attention recently as deep learning has advanced.…”
Section: Introductionmentioning
confidence: 99%