2020
DOI: 10.48550/arxiv.2010.16335
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Calibration-Aided Edge Inference Offloading via Adaptive Model Partitioning of Deep Neural Networks

Abstract: Mobile devices can offload deep neural network (DNN)-based inference to the cloud, overcoming local hardware and energy limitations. However, offloading adds communication delay, thus increasing the overall inference time, and hence it should be used only when needed. An approach to address this problem consists of the use of adaptive model partitioning based on early-exit DNNs. Accordingly, the inference starts at the mobile device, and an intermediate layer estimates the accuracy: If the estimated accuracy i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 11 publications
(16 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?