2021
DOI: 10.1109/jstsp.2021.3058895
|View full text |Cite
|
Sign up to set email alerts
|

RODNet: A Real-Time Radar Object Detection Network Cross-Supervised by Camera-Radar Fused Object 3D Localization

Abstract: Various autonomous or assisted driving strategies have been facilitated through the accurate and reliable perception of the environment around a vehicle. Among the commonly used sensors, radar has usually been considered as a robust and cost-effective solution even in adverse driving scenarios, e.g., weak/strong lighting or bad weather. Instead of considering to fuse the unreliable information from all available sensors, perception from pure radar data becomes a valuable alternative that is worth exploring. Ho… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
58
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 111 publications
(70 citation statements)
references
References 52 publications
0
58
0
Order By: Relevance
“…As we know, deep learning is data-driven, and quality data are crucial for neural network training. At present, the public datasets containing radar information mainly include Nuscence [ 35 ], CRUW [ 36 ] and Oxford Radar Robotcar Dataset [ 37 ]. However, these datasets contain only 2D radar point information or radiofrequency (RF) images of the radar.…”
Section: Resultsmentioning
confidence: 99%
“…As we know, deep learning is data-driven, and quality data are crucial for neural network training. At present, the public datasets containing radar information mainly include Nuscence [ 35 ], CRUW [ 36 ] and Oxford Radar Robotcar Dataset [ 37 ]. However, these datasets contain only 2D radar point information or radiofrequency (RF) images of the radar.…”
Section: Resultsmentioning
confidence: 99%
“…Afterward, conventional 3D transposed convolution and an extra 3D convolutional operation are performed to decode the spatial information from the embedded Spatio-temporal feature Z 3 + Z 3 . Finally, a neck layer is used to project the decoded Spatio-temporal feature to the confidence map with shape × × × × , and followed by applying L-NMS in [15] to retrieve the detection results.…”
Section: Our Dcsn Detection Networkmentioning
confidence: 99%
“…In this manner, the radar object detection can be performed via a single neural network without two-stage or two data sources in the inference phase. In [15], the temporal deformable convolution (TDC) was further proposed to capture the radar features in time effectively. In this way, the performance can be boosted further without increasing inference time significantly.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…There is only a few researches [12,16,17] about how to utilize the rich semantic information in the radar signal for the object detection task. Although those related researches have explored the object detection task on RF data, there is no public benchmark in this area until that piece of work [26] exists, to the best of our knowledge. The most important innovation of that paper is it proposed a cross-model supervision pipeline for annotating object labels on the radio frequency (RF) images by a camera-radar fused (CRF) algorithm automatically and established a public benchmark for radar object detection task.…”
Section: Introductionmentioning
confidence: 99%