2018
DOI: 10.1109/tcad.2018.2857338
|View full text |Cite
|
Sign up to set email alerts
|

Trading-Off Accuracy and Energy of Deep Inference on Embedded Systems: A Co-Design Approach

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 27 publications
(10 citation statements)
references
References 18 publications
0
8
0
Order By: Relevance
“…The core idea is to adapt models during inference based on the complexity of the current task. So far, the model adaption happens at least along four common dimensions: number of layers or cascaded models [66][67][68], number of channels [69][70][71][72], input image resolution [73], and computation precision [74,75].…”
Section: Adaptive Inferencementioning
confidence: 99%
See 1 more Smart Citation
“…The core idea is to adapt models during inference based on the complexity of the current task. So far, the model adaption happens at least along four common dimensions: number of layers or cascaded models [66][67][68], number of channels [69][70][71][72], input image resolution [73], and computation precision [74,75].…”
Section: Adaptive Inferencementioning
confidence: 99%
“…Along the layer dimension, Panda et al [66] propose a conditional deep learning (CDL) network, which can identify the variability in the difficulty of input instances and conditionally activate the deeper layers of the network. Extended from CDL [66], Jayakodi et al [67] propose onthe-fly classifier selection: simple classifiers for easy inputs and complex classifiers for hard inputs. Stamoulis et al [68] propose a systematic approach for hyper-parameter optimization of adaptive CNNs, using Bayesian optimization to determine the number of channels, kernel sizes, and the number of units in the fully connected layers.…”
Section: Adaptive Inferencementioning
confidence: 99%
“…(9) Therefore, in Algorithm 1, we select the next ReRAM design and the fidelity of ReSNA pair that maximizes the information gain per unit cost about the optimal Pareto front based on Equation (9).…”
Section: Selecting Reram Design To Evaluate Via Information Gainmentioning
confidence: 99%
“…However, a key challenge in executing DNN inferencing [9]- [11] on ReRAM-based architecture arises due to nonidealities of ReRAM devices, which can degrade the accuracy of inferencing. Since DNN inferencing involves a sequence of forward computations over DNN layers, errors due to device nonidealities can propagate and accumulate, resulting in incorrect predictions.…”
Section: Introductionmentioning
confidence: 99%
“…Besides, since computation directly translates into energy consumption and IoT devices are usually battery-constrained [9], the high computation demand of training will quickly drain the battery. While existing works [10]- [12] effectively reduce the computation cost of inference by assigning input instances to different classifiers according to the difficulty, the computation cost of training is not reduced.…”
Section: Introductionmentioning
confidence: 99%