2018 51st Annual IEEE/ACM International Symposium on Microarchitecture (MICRO) 2018
DOI: 10.1109/micro.2018.00024
|View full text |Cite
|
Sign up to set email alerts
|

PermDNN: Efficient Compressed DNN Architecture with Permuted Diagonal Matrices

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
68
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 85 publications
(68 citation statements)
references
References 53 publications
0
68
0
Order By: Relevance
“…Computation dataflow. The PE operations in prior sparse CNN implementation [6], [13], [16] are based on vectormatrix and vector-vector matrix multiplications. However, these operations need to re-gather the sparse weights into a new vector or matrix, resulting in the overhead of matching the index between the vector and the matrix.…”
Section: Dataflow and Pe Designmentioning
confidence: 99%
See 3 more Smart Citations
“…Computation dataflow. The PE operations in prior sparse CNN implementation [6], [13], [16] are based on vectormatrix and vector-vector matrix multiplications. However, these operations need to re-gather the sparse weights into a new vector or matrix, resulting in the overhead of matching the index between the vector and the matrix.…”
Section: Dataflow and Pe Designmentioning
confidence: 99%
“…The evolution of DNNs has already piqued interest in hardware acceleration as both DNNs training and inference demand a tremendous amount of computation. As a result, hardware accelerators such as GPUs, FPGAs, and customized ASICs have been employed to accelerate DNNs [4], [5], [6], [7], [8], [9], [10], [11], [12], [13], [14], [15], [16], [17], [18], [19], [20], [21], [22], [23], [24], [25]. However, DNN designers are still hampered by the growing complexity of DNN models.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…Pruning or compressing an already-trained DNN could result in large approximation error [54][55][56][57]. One alternative is to train a sparse DNN.…”
Section: Algorithmic Designmentioning
confidence: 99%