2020 IEEE Real-Time and Embedded Technology and Applications Symposium (RTAS) 2020
DOI: 10.1109/rtas48715.2020.000-8
|View full text |Cite
|
Sign up to set email alerts
|

Real-Time Object Detection System with Multi-Path Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
25
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 41 publications
(30 citation statements)
references
References 35 publications
0
25
0
Order By: Relevance
“…Wang et al [60] navigate the performance-power tradeoff space of mobile SoCs equipped with heterogeneous processors when they perform ML inferences. Heo et al [28] propose an ML inference latency prediction model for GPU and devises multipath neural networks, which enable the runtime to choose which path to take to meet real-time latency constraints. AutoScale [36] is an execution scaling engine that leverages Reinforcement Learning to adaptively determine which platform to pick for performing inference to improve energy efficiency in edge-cloud systems.…”
Section: Inference Task Scheduling For Heterogeneous-platform Edge Devicementioning
confidence: 99%
“…Wang et al [60] navigate the performance-power tradeoff space of mobile SoCs equipped with heterogeneous processors when they perform ML inferences. Heo et al [28] propose an ML inference latency prediction model for GPU and devises multipath neural networks, which enable the runtime to choose which path to take to meet real-time latency constraints. AutoScale [36] is an execution scaling engine that leverages Reinforcement Learning to adaptively determine which platform to pick for performing inference to improve energy efficiency in edge-cloud systems.…”
Section: Inference Task Scheduling For Heterogeneous-platform Edge Devicementioning
confidence: 99%
“…Estimating the worst-case execution time is also discussed in some other works [24], [25]. Besides, there are some works [27], [28] concerning query time estimation in the database context.…”
Section: Related Workmentioning
confidence: 99%
“…Some of these [26], [27], [29] developed (pure) analytical models and assess the validity of the model. Many of the existing studies [20]- [24], [28] build neural-net based learning models and utilize them for deriving estimated time; many others [2], [10], [16]- [19], [25], [30] use tree and linear regression based machine learning models. Some other works [11], [12], [14], [15] use hybrid methods combining these tools-analytical model, machine learning, and deep learning.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…To address the issue, recent works, e.g., [24,25], have investigated how to dynamically skip or add layers to meet timing constraints. Unlike the non-adaptive baseline, we support methodical trade-offs between the inference time and accuracy.…”
Section: Introductionmentioning
confidence: 99%