Proceedings of the 15th ACM Conference on Embedded Network Sensor Systems 2017
DOI: 10.1145/3131672.3131675
|View full text |Cite
|
Sign up to set email alerts
|

DeepIoT

Abstract: Recent advances in deep learning motivate the use of deep neutral networks in sensing applications, but their excessive resource needs on constrained embedded devices remain an important impediment. A recently explored solution space lies in compressing (approximating or simplifying) deep neural networks in some manner before use on the device. We propose a new compression solution, called DeepIoT, that makes two key contributions in that space. First, unlike current solutions geared for compressing speci c ty… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 150 publications
(17 citation statements)
references
References 52 publications
0
7
0
Order By: Relevance
“…Some frameworks for executing neural networks distributed into several IoT devices have been proposed [13], [15], [27], [28], [29], [30]. Sze et al [22] reviewed methods for efficiently executing DNNs, focusing on the inference phase, hardware platforms, and architecture for supporting DNNs.…”
Section: A Machine Learning Frameworkmentioning
confidence: 99%
See 2 more Smart Citations
“…Some frameworks for executing neural networks distributed into several IoT devices have been proposed [13], [15], [27], [28], [29], [30]. Sze et al [22] reviewed methods for efficiently executing DNNs, focusing on the inference phase, hardware platforms, and architecture for supporting DNNs.…”
Section: A Machine Learning Frameworkmentioning
confidence: 99%
“…DeepIoT [15] compresses CNNs, fully connected neural networks, and Recurrent Neural Networks by extracting redundant neurons. This compression can significantly reduce the DNN size, execution time, and energy consumption without loss of accuracy.…”
Section: A Machine Learning Frameworkmentioning
confidence: 99%
See 1 more Smart Citation
“…Its downside, common to many DL compression techniques, is a permanent decrease in inference accuracy (of ≈ 5%). On the pruning front, PatDNN enables real-time inference using large-scale DL models (e.g., VGG-16, ResNet-50) on mobile devices by harnessing pattern-based model pruning [33], while DeepIoT [48] uses reinforcement learning to guide the pruning process. Both solutions lead to significant model size reductions (90% to 98.9% in case of DeepIoT) and speedups (up to 44.5× in case of PatDNN) with no inference accuracy degradation in certain settings, demonstrating vast opportunities for mobile DL optimisation.…”
Section: Related Workmentioning
confidence: 99%
“…Firstly, it has become common for distributed applications to be configured for use on mobile [26,30,34] or embedded [10] devices with on the order of GBs of memory and storage, but they are both very energyhungry and expensive at the scale of our application. Another approach is to reduce the required memory and computation by simplifying the models used, for instance in deep neural nets [24,42,43]. However, in cases like ours where mobile phones are too expensive and power hungry and the problem can't be adequately simplified, the use of a co-processor comes to mind.…”
Section: Related Workmentioning
confidence: 99%