2021
DOI: 10.48550/arxiv.2112.11413
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Offloading Algorithms for Maximizing Inference Accuracy on Edge Device Under a Time Constraint

Abstract: With the emergence of edge computing, the problem of offloading jobs between an Edge Device (ED) and an Edge Server (ES) received significant attention in the past. Motivated by the fact that an increasing number of applications are using Machine Learning (ML) inference from the data samples collected at the EDs, we study the problem of offloading inference jobs by considering the following novel aspects: 1) in contrast to a typical computational job, the processing time of an inference job depends on the size… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 23 publications
0
2
0
Order By: Relevance
“…The authors in [8] proposed AM R 2 , a scheduling scheme based on LP-Relaxation and rounding which considers all possible cases of scheduling two inference tasks between the edge device and the edge server. They relax the problem's constraint to take fractional values and then perform rounding to get the result.…”
Section: A Inference Task Scheduling and Offloadingmentioning
confidence: 99%
See 1 more Smart Citation
“…The authors in [8] proposed AM R 2 , a scheduling scheme based on LP-Relaxation and rounding which considers all possible cases of scheduling two inference tasks between the edge device and the edge server. They relax the problem's constraint to take fractional values and then perform rounding to get the result.…”
Section: A Inference Task Scheduling and Offloadingmentioning
confidence: 99%
“…This flexibility allows edge nodes to deploy multiple local inference models, each optimized for different accuracy levels. [8] Considering an edge computing system empowered by AI models in which edge devices receive sensed data from the en-vironment or from user interactions (see Fig 1). These devices are equipped with pretrained inference models which can be used to perform inference on data locally.…”
Section: Introductionmentioning
confidence: 99%