2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) 2019
DOI: 10.1109/iccvw.2019.00114
|View full text |Cite
|
Sign up to set email alerts
|

Monocular 3D Object Detection with Pseudo-LiDAR Point Cloud

Abstract: Monocular 3D scene understanding tasks, such as object size estimation, heading angle estimation and 3D localization, is challenging. Successful modern day methods for 3D scene understanding require the use of a 3D sensor. On the other hand, single image based methods have significantly worse performance. In this work, we aim at bridging the performance gap between 3D sensing and 2D sensing for 3D object detection by enhancing LiDAR-based algorithms to work with single image input. Specifically, we perform mon… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
142
1

Year Published

2019
2019
2020
2020

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 246 publications
(159 citation statements)
references
References 74 publications
0
142
1
Order By: Relevance
“…To evaluate the true potential and fair comparison with state-of-the-art systems, MODT is evaluated on the KITTI datasets with ground truths [44], based on the MOT16 evaluation metrics proposed in [41]. Whereas, the MOT component is independently evaluated using off-the-shelf 3D object detector [45], [46] against the 3D MOT evaluation extension [19].…”
Section: Discussionmentioning
confidence: 99%
“…To evaluate the true potential and fair comparison with state-of-the-art systems, MODT is evaluated on the KITTI datasets with ground truths [44], based on the MOT16 evaluation metrics proposed in [41]. Whereas, the MOT component is independently evaluated using off-the-shelf 3D object detector [45], [46] against the 3D MOT evaluation extension [19].…”
Section: Discussionmentioning
confidence: 99%
“…Table 2 shows the localization results at a short distance, Table 3 shows the localization results at a medium distance, and Table 4 shows the localization results at a long distance. We use the formula (1) to calculate the location accuracy, depth p represents the predicted depth and depth t represents the true depth, "|*|" represents the absolute value calculation. In order to get the effective accuracy, that is, the object predicted depth is comparable with the real depth, the predicted depth should not be more than twice the real depth, otherwise it is meaningless.…”
Section: Comparison With Unsupervised Depth Estimation Based On Deep mentioning
confidence: 99%
“…Hardware-assisted methods rely on non-visual sensors. Weng et al proposed a monocular three-dimensional object detection algorithm based on pseudo-laser point cloud [1], Yang et al proposed and implemented a monocular vision SLAM-based UAV autonomous landing system with fusion optical equipment and image recognition technology [2], Wei et al proposed a monocular image obstacle detection algorithm that combines laser radar sparse point clouds and camera natural image density clustering [3], mainly focused on obstacle detection, indoor localization, indoor path planning and other application scenarios [4][5][6][7][8]. Traditional geometric methods are generally based on binocular vision.…”
Section: Introductionmentioning
confidence: 99%
“…However, in order to predict future trajectory reliably, these methods rely on accurate human trajectory history. This usually involves multi-people detection [23], [24], [25], [26] and tracking, and thus has two major disadvantages: (1) Data association is very challenging in crowded scenarios. Small mistakes can be misinterpreted to be very large changes of velocity, resulting in very bad trajectory estimate; (2) Robust multi-person tracking from mobile platform is often timeconsuming.…”
Section: Related Workmentioning
confidence: 99%