2020
DOI: 10.1109/jssc.2020.3005786
|View full text |Cite
|
Sign up to set email alerts
|

An Energy-Efficient Deep Convolutional Neural Network Training Accelerator for In Situ Personalization on Smart Devices

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 37 publications
(12 citation statements)
references
References 18 publications
0
9
0
Order By: Relevance
“…All the coded indexes are stored in the Para-Buffer, and only the indexes need to be loaded into the AF and pooling process engines in the BP phase. The index coding reduces 50% storage compares to the design proposed in Ref [12].…”
Section: A Computation and Dataflowmentioning
confidence: 94%
See 2 more Smart Citations
“…All the coded indexes are stored in the Para-Buffer, and only the indexes need to be loaded into the AF and pooling process engines in the BP phase. The index coding reduces 50% storage compares to the design proposed in Ref [12].…”
Section: A Computation and Dataflowmentioning
confidence: 94%
“…Our processor achieves an energy efficiency of 2.44 TOPS/W for inference and 1.36 TOPS/W for training. Compare to the latest training dedicated processor Ref [12], our processor yields a 1.32х energy efficiency improvement for training. Compare to the latest inference/training processor Ref [11], our design achieves 2.1х energy efficiency improvement for training and 1.09х for inference.…”
Section: B Energy Efficiency Comparisonmentioning
confidence: 96%
See 1 more Smart Citation
“…Previous work on accelerating the DNN training has focused on leveraging the sparsity present in weights and activa- tions [11], [33], [44], [45]. TensorDash [33] accelerates the DNN training process while achieving higher energy efficiency via eliminating the ineffectual operations resulted from the sparse input data.…”
Section: Accelerators For Dnn Trainingmentioning
confidence: 99%
“…In terms of computing, a low-power CPU will be used for initiating data movement, performing data preprocessing tasks (normalization and dimensionality reduction), and invoking a specialized MLP accelerator. The accelerator is assumed to be able to support high-throughput, low-latency inference as well as on-device training of MLPs, similar to some of the recent advances (Choi et al, 2020). This paper introduces lightweight MLP-LIBS-ADAPT for portable and remote LIBS systems, which can also adapt to any domain shift in a semi-supervised manner.…”
Section: Accelerator Design For Libsmentioning
confidence: 99%