2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.01173
|View full text |Cite
|
Sign up to set email alerts
|

Train in Germany, Test in the USA: Making 3D Object Detectors Generalize

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
164
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 122 publications
(168 citation statements)
references
References 47 publications
4
164
0
Order By: Relevance
“…Evaluation Metric. We evaluate SEE on the "Car" category in the KITTI validation dataset, similar to other UDA methods [56,64]. We follow the official KITTI evaluation metric and report the average precision (AP) over 40 recall positions at 0.7 and 0.5 IoU thresholds for both BEV and 3D IoUs.…”
Section: See Results On Public Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…Evaluation Metric. We evaluate SEE on the "Car" category in the KITTI validation dataset, similar to other UDA methods [56,64]. We follow the official KITTI evaluation metric and report the average precision (AP) over 40 recall positions at 0.7 and 0.5 IoU thresholds for both BEV and 3D IoUs.…”
Section: See Results On Public Datasetsmentioning
confidence: 99%
“…For the specific task of labelled source domain to unlabelled target domain for 3D object detection across distinct lidar scan patterns, research has been more sparse. Wang et al [56] proposed a semi-supervised approach using object-size statistics of the target domain to resize training samples in the labelled source domain. A popular approach is the use of self-training [43,63,64,67] with a focus on generating quality pseudo-labels using temporal information [43,67] or an IoU scoring criterion for historical pseudo-labels [63,64].…”
Section: Related Workmentioning
confidence: 99%
“…Besides, note that the camera parameters of the images on the KITTI test set are different from these of the training/validation set, and the good performance on the test set suggests the proposed method can also generalize to different camera parameters. However, generalizing to the new scenes with different statistical characteristics is a hard task for existing 3D detectors Wang et al, 2020b), including the image-based models and LiDAR-based models, and deserves further investigation by future works. We also argue that the proposed method can generalize to the new scenes better than other monocular models because ours model learns the stronger features from the teacher net.…”
Section: A5 Generalization Of the Proposed Methodsmentioning
confidence: 99%
“…Note that during these 3D object detection experiments, we only make use of angle estimators pre-trained on virtual data. Our method can therefore be considered as fully self-supervised without the use human annotations, which is in contrast to DA-based methods as in [41].…”
Section: Methodsmentioning
confidence: 99%
“…Many of these methods rely either on synthetic data or make use of adversarial feature learning [40]. Recently, the DA task for full 3D object detection has also been considered in [41] for one of the first times. 4) 3D bounding box optimization: We use a 3D bounding box optimization process to obtain a 3D box from a 2D detector and its yaw angle estimate, based on geometrical constraints.…”
Section: ) Monocular Vehicle Orientation Estimationmentioning
confidence: 99%