2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2019
DOI: 10.1109/cvprw.2019.00162
|View full text |Cite
|
Sign up to set email alerts
|

Sensor Fusion for Joint 3D Object Detection and Semantic Segmentation

Abstract: In this paper, we present an extension to LaserNet, an efficient and state-of-the-art LiDAR based 3D object detector. We propose a method for fusing image data with the LiDAR data and show that this sensor fusion method improves the detection performance of the model especially at long ranges. The addition of image data is straightforward and does not require image labels. Furthermore, we expand the capabilities of the model to perform 3D semantic segmentation in addition to 3D object detection. On a large ben… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

2
81
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 123 publications
(85 citation statements)
references
References 32 publications
2
81
0
Order By: Relevance
“…Our proposed method outperforms the previous methods that only utilize LiDAR data. Furthermore, by simply changing the loss function, we observe a similar gain in performance as adding an additional sensing modality (see LaserNet++ [11]). Fig.…”
Section: A Detection Evaluationmentioning
confidence: 72%
See 1 more Smart Citation
“…Our proposed method outperforms the previous methods that only utilize LiDAR data. Furthermore, by simply changing the loss function, we observe a similar gain in performance as adding an additional sensing modality (see LaserNet++ [11]). Fig.…”
Section: A Detection Evaluationmentioning
confidence: 72%
“…To accomplish this task, autonomous vehicles are equipped with various sensors including cameras and LiDARs. A wealth of deep learning based approaches have been proposed to perform 3D object detection using these sensors [1], [2], [3], [4], [5], [6], [7], [8], [9], [10], [11]. Given the limited sensory information, it is unrealistic to expect any detector to flawlessly classify and localize every actor in all situations.…”
Section: Introductionmentioning
confidence: 99%
“…Methods such as [20,21] propose the idea of multi-sensor fusion networks [20,21] to increase the model accuracy, but despite high accuracy, these methods are computationally expensive. To tackle the sensor-fusion computational problem, this [22] proposed an earlyfusion method to fuse both camera and LiDAR with only one backbone, attaining a good balance between accuracy and efficiency.…”
Section: D Object Detectionmentioning
confidence: 99%
“…There are various types of sensor fusion depending on where the data are projected. For studies using a bird’s eye-view (BEV), there is the fusion method of cameras and LiDAR by using the deep learning technique after calibration [ 1 ]. In many studies, fusion is performed by projecting the point cloud of LiDAR onto the image space.…”
Section: Introductionmentioning
confidence: 99%