Proceedings of the International Conference on Industrial Control Network and System Engineering Research 2019
DOI: 10.1145/3333581.3333589
|View full text |Cite
|
Sign up to set email alerts
|

Layer-by-layer Quantization Method for Neural Network Parameters

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 6 publications
0
1
0
Order By: Relevance
“…Through parameter quantization, the resolution of the parameters can be reduced to 16-bit, 8-bit, 4-bit, and even 1-bit with little loss of the accuracy in some tasks. [26][27][28] Prakosa et al 29 adopted the K-D method to improve the performance of the pruned network. Blakeney et al 30 proposed a parallel block-wise K-D method to compress the deep neural networks.…”
Section: Introductionmentioning
confidence: 99%
“…Through parameter quantization, the resolution of the parameters can be reduced to 16-bit, 8-bit, 4-bit, and even 1-bit with little loss of the accuracy in some tasks. [26][27][28] Prakosa et al 29 adopted the K-D method to improve the performance of the pruned network. Blakeney et al 30 proposed a parallel block-wise K-D method to compress the deep neural networks.…”
Section: Introductionmentioning
confidence: 99%