2021 Symposium on VLSI Circuits 2021
DOI: 10.23919/vlsicircuits52068.2021.9492420
|View full text |Cite
|
Sign up to set email alerts
|

A 28nm 276.55TFLOPS/W Sparse Deep-Neural-Network Training Processor with Implicit Redundancy Speculation and Batch Normalization Reformulation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 17 publications
(11 citation statements)
references
References 0 publications
0
10
0
Order By: Relevance
“…Output sparsity exploitation during the WG stage has big benefits thanks to both useless computation avoidance and memory access removal. For this reason, recent energy-efficient training processors [26,31,51,78] supported triple sparsity exploitation by combining iterative pruning.…”
Section: Pruning-aware Output Zero Skipping During the Wgmentioning
confidence: 99%
See 4 more Smart Citations
“…Output sparsity exploitation during the WG stage has big benefits thanks to both useless computation avoidance and memory access removal. For this reason, recent energy-efficient training processors [26,31,51,78] supported triple sparsity exploitation by combining iterative pruning.…”
Section: Pruning-aware Output Zero Skipping During the Wgmentioning
confidence: 99%
“…[29] [26,31,51,78] Sparsity Exploitation [17][18][19] was proposed to unify the data representation method of both input operand and accumulation. Flexpoint [55] tried to substitute FP with FXP representation using a shared exponent management algorithm together for simplification of MAC design, but it failed to reduce the required bit-precision to less than 16-bit.…”
Section: A New Number Representationmentioning
confidence: 99%
See 3 more Smart Citations