2020 53rd Annual IEEE/ACM International Symposium on Microarchitecture (MICRO) 2020
DOI: 10.1109/micro50266.2020.00069
|View full text |Cite
|
Sign up to set email alerts
|

TensorDash: Exploiting Sparsity to Accelerate Deep Neural Network Training

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
25
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 52 publications
(26 citation statements)
references
References 35 publications
1
25
0
Order By: Relevance
“…To demonstrate the effectiveness of Shift-BNN, we compare it with three training accelerators: Firstly, since Shift-BNN adopts RC-mapping as the fundamental design strategy, we compare it with the RC-accelerator that adopts RC-mapping strategy but without LFSR reversion technique. Secondly, since MN-mapping is commonly used in existing DNN training accelerators [39,64], we employ an MN-accelerator that adopts MNmapping strategy without LFSR reversion technique as the baseline accelerator for generality, which is also used for our preliminary investigation in Sec.3. Thirdly, to verify the analysis about design alternatives (see Sec.5), we further test the effectiveness of our LFSR reversion strategy on MN-accelerator by comparing with an MN-Shift-accelerator that adopts both MN-mapping strategy and LFSR reversion technique.…”
Section: Evaluation 71 Experimental Methodologymentioning
confidence: 99%
See 1 more Smart Citation
“…To demonstrate the effectiveness of Shift-BNN, we compare it with three training accelerators: Firstly, since Shift-BNN adopts RC-mapping as the fundamental design strategy, we compare it with the RC-accelerator that adopts RC-mapping strategy but without LFSR reversion technique. Secondly, since MN-mapping is commonly used in existing DNN training accelerators [39,64], we employ an MN-accelerator that adopts MNmapping strategy without LFSR reversion technique as the baseline accelerator for generality, which is also used for our preliminary investigation in Sec.3. Thirdly, to verify the analysis about design alternatives (see Sec.5), we further test the effectiveness of our LFSR reversion strategy on MN-accelerator by comparing with an MN-Shift-accelerator that adopts both MN-mapping strategy and LFSR reversion technique.…”
Section: Evaluation 71 Experimental Methodologymentioning
confidence: 99%
“…DNN training optimization has been extensively studied [39,44,49,60,64]. For example, eager pruning [64] and Procrustes [60] exploit the weight sparsity during the training stage by leveraging aggressive pruning algorithms and develop customized hardware to improve the performance.…”
Section: Related Workmentioning
confidence: 99%
“…Previous work on accelerating the DNN training has focused on leveraging the sparsity present in weights and activa- tions [11], [33], [44], [45]. TensorDash [33] accelerates the DNN training process while achieving higher energy efficiency via eliminating the ineffectual operations resulted from the sparse input data.…”
Section: Accelerators For Dnn Trainingmentioning
confidence: 99%
“…Previous work on accelerating the DNN training has focused on leveraging the sparsity present in weights and activa- tions [11], [33], [44], [45]. TensorDash [33] accelerates the DNN training process while achieving higher energy efficiency via eliminating the ineffectual operations resulted from the sparse input data. Eager Pruning [45] and Procrustes [44] improve DNN training efficiency by co-designing the training algorithm with the target hardware platform ("hardware-aware training").…”
Section: Accelerators For Dnn Trainingmentioning
confidence: 99%
“…Since the sparsification is dynamically changing most of the designs are not efficient to leverage the sparse computation in HASI. In order to address the challenges of dynamic sparsification in HASI we developed a hardware-software codesigned accelerator for a dynamically sparsified model, based on the TensorDash [23] architecture. We call our design a Dynamic sparsified CNN (DySCNN) accelerator.…”
Section: A Noisy Sparsificationmentioning
confidence: 99%