2020
DOI: 10.1109/jiot.2020.2972000
|View full text |Cite
|
Sign up to set email alerts
|

Toward Collaborative Inferencing of Deep Neural Networks on Internet-of-Things Devices

Abstract: With recent advancements in deep neural networks (DNNs), we are able to solve traditionally challenging problems. Since DNNs are compute intensive, consumers, to deploy a service, need to rely on expensive and scarce compute resources in the cloud. This approach, in addition to its dependability on high-quality network infrastructure and data centers, raises new privacy concerns. These challenges may limit DNN-based applications, so many researchers have tried optimize DNNs for local and in-edge execution. How… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
51
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 74 publications
(54 citation statements)
references
References 31 publications
0
51
0
Order By: Relevance
“…applying different approaches for model-parallelism. The results of Hadidi et al [2019] show that the collaborative network is enhanced by creating a distributed processing pipeline. Differently from these works, we are interested in the partitioning of DNN models between edge and cloud.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…applying different approaches for model-parallelism. The results of Hadidi et al [2019] show that the collaborative network is enhanced by creating a distributed processing pipeline. Differently from these works, we are interested in the partitioning of DNN models between edge and cloud.…”
Section: Related Workmentioning
confidence: 99%
“…The results show that their proposed method is able to reduce the network communication costs without harming accuracy. A similar approach is proposed by Hadidi et al [2019], which aggregates existing computing power of edge devices in an local network environment by creating a collaborative network. In this scenario, edge devices cooperate to make inferences by…”
Section: Related Workmentioning
confidence: 99%
“…Collaborative perception [18] pipelines the computation by partitioning a DNN model and distributing the partitioned blocks to multiple edge devices. Follow-up work in [17] introduces model parallelism methods for both dense and convolution layers in order to address the memory overhead due to model replication. DeepThings [51] reduces the communication cost by fusing the early convolution layers and parallelizing these layers in multiple devices.…”
Section: Related Workmentioning
confidence: 99%
“…In response to the aforementioned concerns, efforts have been made to push machine learning inference from the cloud to the edge. Edge processing has the benefit of keeping data closer to its source to provide real-time responses while protecting the privacy of the end-user [17,29]. Unfortunately, edge computing for machine learning workloads faces challenges of its own.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation