2021
DOI: 10.48550/arxiv.2103.04505
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Split Computing and Early Exiting for Deep Learning Applications: Survey and Research Challenges

Abstract: Mobile devices such as smartphones and autonomous vehicles increasingly rely on deep neural networks (DNNs) to execute complex inference tasks such as image classification and speech recognition, among others. However, continuously executing the entire DNN on the mobile device can quickly deplete its battery. Although task offloading to edge devices may decrease the mobile device's computational burden, erratic patterns in channel quality, network and edge server load can lead to a significant delay in task ex… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
19
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 12 publications
(19 citation statements)
references
References 88 publications
(228 reference statements)
0
19
0
Order By: Relevance
“…Split Computing (SC) [48] is a framework that divides the DNN model into head and tail models, which are executed in the edge device and server, respectively. SC is attractive when compressed models for edge devices cannot achieve the same level of accuracy as their full counterpart models.…”
Section: Split Computationmentioning
confidence: 99%
“…Split Computing (SC) [48] is a framework that divides the DNN model into head and tail models, which are executed in the edge device and server, respectively. SC is attractive when compressed models for edge devices cannot achieve the same level of accuracy as their full counterpart models.…”
Section: Split Computationmentioning
confidence: 99%
“…We use image data with relatively high resolution, including ImageNet [52], COCO [53], and PASCAL VOC datasets [54]. As pointed out in [55], split computing is mainly beneficial for supervised tasks involving high-resolution images e.g., 224 × 224 pixels or larger. For smaller data, either local processing or full offloading would be more suitable.…”
Section: Choice Of Datasetsmentioning
confidence: 99%
“…Other surveys. There have been certain previous surveys touching on the topic of early-exiting, either only briefly discussing it from the standpoint of dynamic inference networks [21] or combining it with offloading [54]. To the best of our knowledge, this is the first study that primarily focuses on early-exit networks and their design trade-offs across tasks, modalities and target hardware.…”
Section: Adaptive Inference Landscapementioning
confidence: 99%