2021 IEEE Winter Conference on Applications of Computer Vision (WACV) 2021
DOI: 10.1109/wacv48630.2021.00232
|View full text |Cite
|
Sign up to set email alerts
|

Generalized Object Detection on Fisheye Cameras for Autonomous Driving: Dataset, Representations and Baseline

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
29
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
3

Relationship

4
5

Authors

Journals

citations
Cited by 63 publications
(40 citation statements)
references
References 21 publications
0
29
0
Order By: Relevance
“…The standard bounding box representation fails in fisheye cameras due to heavy radial distortion, particularly in the periphery. In concurrent work [15], we explored different output representations for fisheye images, including oriented bounding boxes, curved boxes, ellipses, and polygons. We have integrated this model in our MTL framework where we use a 24-sided polygon representation for object detection.…”
Section: B Generalized Object Detectionmentioning
confidence: 99%
“…The standard bounding box representation fails in fisheye cameras due to heavy radial distortion, particularly in the periphery. In concurrent work [15], we explored different output representations for fisheye images, including oriented bounding boxes, curved boxes, ellipses, and polygons. We have integrated this model in our MTL framework where we use a 24-sided polygon representation for object detection.…”
Section: B Generalized Object Detectionmentioning
confidence: 99%
“…Motion segmentation employs two frames and predicts either a binary moving or static mask and employs Lovasz-Softmax [19], and Focal [20] loss for managing class imbalance instead of the cross-entropy loss. For object detection, we make use of YOLOv3 loss and add IoU loss using segmentation mask [21].…”
Section: Supervised Trainingmentioning
confidence: 99%
“…There is relatively less work on using convolutional neural networks for fisheye cameras. Recently, it has been explored for various tasks such as object detection [13], soiling detection [14], depth prediction [15], [16], [17], [18], SLAM [19] and in general for multi-task outputs [20]. As we shall see later, of particular interest for chargepad detection is the joint learning of object, depth, and semantic segmentation, which we shall deploy for the refinement of auto-annotation for online learning.…”
Section: Front Rightmentioning
confidence: 99%
“…Motion segmentation employs two frames and predicts either a binary moving or static mask and also employs Lovasz-Softmax and Focal [29] loss for managing class imbalance instead of the cross-entropy loss. For object detection, we make use of YOLOv3 loss [30] and add IoU loss using segmentation mask [13].…”
Section: B Near-field Perception Stack Through Multi-task Learning (Mtl)mentioning
confidence: 99%