2021
DOI: 10.1016/bs.adcom.2020.07.002
|View full text |Cite
|
Sign up to set email alerts
|

Energy-efficient deep learning inference on edge devices

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
24
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

3
6

Authors

Journals

citations
Cited by 26 publications
(24 citation statements)
references
References 31 publications
0
24
0
Order By: Relevance
“…Machine Learning (ML) plays an increasingly important role in many Internet of Things (IoT) applications, ranging from computer vision to time-series processing [7,9,28,31]. Edge computing, as a paradigm to host data-analytics as close as possible to end devices, may offer several advantages compared to the standard cloud-centric approach.…”
Section: Introductionmentioning
confidence: 99%
“…Machine Learning (ML) plays an increasingly important role in many Internet of Things (IoT) applications, ranging from computer vision to time-series processing [7,9,28,31]. Edge computing, as a paradigm to host data-analytics as close as possible to end devices, may offer several advantages compared to the standard cloud-centric approach.…”
Section: Introductionmentioning
confidence: 99%
“…However, the high specialization of the neural accelerators makes them potentially much more energyefficient, a critical parameter in embedded systems. These processors can be found under different names such as Tensor Processing Unit (TPU) [33], Neural Processing Unit (NPU) [34], or Vision Processing Unit (VPU) [35].…”
Section: E Deep Learning Neural Accelerators For the Edgementioning
confidence: 99%
“…In practice, this result is obtained masking different slices of each layer's weights with binary parameters, so that the slices multiplied with a 0 are effectively eliminated from the layer. The continuous relaxation of the binary mask is then optimized, similarly to the architectural weights in a super-net DNAS, with the objective of reducing the network complexity, by eliminating unimportant parts of each layer (in that, this approach is similar to a structured pruning [5]). The usage of masks introduces a minimum overhead with respect to a normal training of the seed [20], reducing the search time and memory requirements significantly compared to super-net approaches, and representing a further step towards lightweight NAS.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Deep Learning (DL) is at the core of many modern computing applications, such as computer vision [1], sound classification [2], bio-signal analysis [3], predictive maintenance [4], etc. While DL models have been traditionally deployed on powerful cloud-based servers, evidence exists about the potential advantages of an implementation at-theedge [5]. Edge computing could improve privacy and reduce the energy consumption at the distributed system level, by replacing the energy hungry wireless transmission of raw data with more efficient local computations and transmission of aggregated outputs [6].…”
Section: Introductionmentioning
confidence: 99%