2020
DOI: 10.1007/978-3-030-58526-6_16
|View full text |Cite
|
Sign up to set email alerts
|

Differentiable Joint Pruning and Quantization for Hardware Efficiency

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

2
36
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 46 publications
(41 citation statements)
references
References 12 publications
2
36
0
Order By: Relevance
“…Experimental results show that, our method achieves a substantially higher compression rate than recent works, while maintains comparable or better performance. For instance, our method achieves higher classification accuracy and compactness than the recent DJPQ [38]. It gets similar performance with the recent DHP [22] on image super-resolution, meanwhile saves about 50% computation.…”
mentioning
confidence: 60%
See 3 more Smart Citations
“…Experimental results show that, our method achieves a substantially higher compression rate than recent works, while maintains comparable or better performance. For instance, our method achieves higher classification accuracy and compactness than the recent DJPQ [38]. It gets similar performance with the recent DHP [22] on image super-resolution, meanwhile saves about 50% computation.…”
mentioning
confidence: 60%
“…However, it is hard to implement unstructured pruning with typical hardware. A recent work DJPQ [38] is proposed combining pruning method VIB-net [4,34] and quantization method DQ [35]. Our work shares certain similarity with DJPQ [38] in that, it also leverages DQ [35] for end-to-end training.…”
Section: Related Workmentioning
confidence: 93%
See 2 more Smart Citations
“…Following [9,19,28], OPs is the sum of low-bit operations and floating-point operations, i.e., for 𝑀-bit networks, OPs = BOPs/64…”
Section: Comparison With Existing Sr Quantization Modelsmentioning
confidence: 99%