2021
DOI: 10.48550/arxiv.2112.12252
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Leveraging Synthetic Data in Object Detection on Unmanned Aerial Vehicles

Abstract: Acquiring data to train deep learning-based object detectors on Unmanned Aerial Vehicles (UAVs) is expensive, time-consuming and may even be prohibited by law in specific environments. On the other hand, synthetic data is fast and cheap to access. In this work, we explore the potential use of synthetic data in object detection from UAVs across various application environments. For that, we extend the open-source framework DeepGTAV to work for UAV scenarios. We capture various largescale high-resolution synthet… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(10 citation statements)
references
References 31 publications
0
10
0
Order By: Relevance
“…This result is in line with the conclusions of previous works addressing the problem of a so-called “reality” gap, existing in the case of using both real (experimental) and “synthetic” data for deep neural network training. Namely, the results obtained during the application of convolutional neural networks in the field of computer vision allow one to conclude that the optimal ratio of real data to “synthetic” data in any deep learning task could be about 5–20% to 80–95% (for a more comprehensive review we refer the reader to Section S6 in the Supporting Information). Since our current experimental data set contains 214 PIs with 607 values of their T g values, one may assume that starting with about 214 real samples at the lowest possible real-to-synthetic data ratio 5% to 95% we will obtain just about 5000 samples in total effective real-and-synthetic training set.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…This result is in line with the conclusions of previous works addressing the problem of a so-called “reality” gap, existing in the case of using both real (experimental) and “synthetic” data for deep neural network training. Namely, the results obtained during the application of convolutional neural networks in the field of computer vision allow one to conclude that the optimal ratio of real data to “synthetic” data in any deep learning task could be about 5–20% to 80–95% (for a more comprehensive review we refer the reader to Section S6 in the Supporting Information). Since our current experimental data set contains 214 PIs with 607 values of their T g values, one may assume that starting with about 214 real samples at the lowest possible real-to-synthetic data ratio 5% to 95% we will obtain just about 5000 samples in total effective real-and-synthetic training set.…”
Section: Resultsmentioning
confidence: 99%
“…In addition, we address the “reality gap” problem that arises from the differences in the size of the experimental and “synthetic” data sets influencing the GCNN training quality. This problem has been first explored in the area of deep learning in computer vision tasks: “synthetic datasets” may contain rather huge values with fail-proof ground-truth labeling; a large amount of data could significantly decrease the model performance on real data. Regarding this issue, we derive important estimates about the necessary amount of data on pretraining stage for the effective development of GCNN to predict polymer properties.…”
Section: Introductionmentioning
confidence: 99%
“…For Video Anomaly Detection, we fall back to the Open-Water data set of the Maritime Anomaly Detection Benchmark [17] since it is the only one featuring metadata.…”
Section: Results and Analysismentioning
confidence: 99%
“…We apply the same memory map from before on both UAVs' object detectors' output, but this time, we average their memory maps in the overlapping region. For that, we take an EfficientDet-D0 trained on POG [17] and test it on the following data. We capture four minutes of footage in a similar environment as POG from the viewpoint of two UAVs, a DJI Matrice 210 and a DJI Mavic 2 Pro, denoted 2AVs.…”
Section: Cooperative Detection Via Multiple Uavsmentioning
confidence: 99%
“…Leveraging Synthetic Data in Object Detection on Unmanned Aerial Vehicles [6] RetinaNet feature pyramid network (FPN) backbone…”
mentioning
confidence: 99%