2022
DOI: 10.3390/app122010619
|View full text |Cite
|
Sign up to set email alerts
|

Partitioning DNNs for Optimizing Distributed Inference Performance on Cooperative Edge Devices: A Genetic Algorithm Approach

Abstract: To fully unleash the potential of edge devices, it is popular to cut a neural network into multiple pieces and distribute them among available edge devices to perform inference cooperatively. Up to now, the problem of partitioning a deep neural network (DNN), which can result in the optimal distributed inferencing performance, has not been adequately addressed. This paper proposes a novel layer-based DNN partitioning approach to obtain an optimal distributed deployment solution. In order to ensure the applicab… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 41 publications
(54 reference statements)
0
3
0
Order By: Relevance
“…In turn, the camera will run the inference task assigned to it and send the corresponding output to the smartphone, which performs the second exit branch and obtains corresponding identification results. There are more similar application scenarios, such as distributed fall detection [ 8 ] and traffic prediction [ 9 ].…”
Section: Introductionmentioning
confidence: 99%
“…In turn, the camera will run the inference task assigned to it and send the corresponding output to the smartphone, which performs the second exit branch and obtains corresponding identification results. There are more similar application scenarios, such as distributed fall detection [ 8 ] and traffic prediction [ 9 ].…”
Section: Introductionmentioning
confidence: 99%
“…Deploying DNN models on edge devices (e.g., embedded systems) presents various challenges, including the limited computational power and memory of edge devices, which can often prevent the deployment of large DNN models entirely. For example, Convolution Neural Networks (CNNs), another type of DNN, can be large and computationally intensive, which makes it challenging to deploy the entire CNN model on a single core edge device [6]. Traditionally, machine learning-based CPS applications were run sequentially on a singlecore processor or device [7].…”
Section: Introductionmentioning
confidence: 99%
“…For example, Chuang Hu et al [12] proposed a min-cut-based algorithm to partition and offload DNNs in both edge and cloud environments. Similarly, cloud and edge-assisted approaches [6,13] divide the DNNs into two parts to offer local and remote computation. However, previous studies have not addressed the ideal number of partitions for the DNN model concerning layer dependencies, device computing power, and communication latency.…”
Section: Introductionmentioning
confidence: 99%