2020 57th ACM/IEEE Design Automation Conference (DAC) 2020
DOI: 10.1109/dac18072.2020.9218746
|View full text |Cite
|
Sign up to set email alerts
|

Q-CapsNets: A Specialized Framework for Quantizing Capsule Networks

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
14
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
2
2
2

Relationship

2
4

Authors

Journals

citations
Cited by 14 publications
(14 citation statements)
references
References 12 publications
0
14
0
Order By: Relevance
“…In particular, in [175], it is stated that the bitwidth used for the weights can decrease approaching the last layers of the NN, while the bitwidth of the activations remains more or less constant. Following these ideas, Q-CapsNets [176] analyzes the layer-wise quantization capabilities of weights and activations of CapsNets, with a cross-layer optimization of the bitwidth and a fine-grained tuning for the dynamic routing operations. Finding the optimal bitwidth for each layer of a DNN is a complex task.…”
Section: F Full Precision Vs Quantized Implementationsmentioning
confidence: 99%
“…In particular, in [175], it is stated that the bitwidth used for the weights can decrease approaching the last layers of the NN, while the bitwidth of the activations remains more or less constant. Following these ideas, Q-CapsNets [176] analyzes the layer-wise quantization capabilities of weights and activations of CapsNets, with a cross-layer optimization of the bitwidth and a fine-grained tuning for the dynamic routing operations. Finding the optimal bitwidth for each layer of a DNN is a complex task.…”
Section: F Full Precision Vs Quantized Implementationsmentioning
confidence: 99%
“…The edge platforms typically have limited memory and power/energy budgets, hence small-sized DNN models with limited number of operations are desired for Edge AI applications. Model compression techniques such as pruning (i.e., structured [15] [16] or unstructured [17]- [19]) and quantization [19]- [22] are considered to be highly effective for reducing the memory footprint of the models as well as for reducing the number of computations required per inference. Structured pruning [15] can achieve about 4x weight memory compression, while class-blind unstructured pruning (i.e., PruNet [18]) achieves up to 190x memory compression.…”
Section: A Optimizations For Dnn Modelsmentioning
confidence: 99%
“…For instance, quantization in the Deep Compression [19] improves the compression rate by about 3x for the AlexNet and the VGG-16 models. The Q-CapsNets framework [22] shows that quantization is highly effective for complex DNNs such as CapsNets as well. It reduces the memory requirement of the CapsNet [14] by 6.2x with a negligible accuracy degradation of 0.15%.…”
Section: A Optimizations For Dnn Modelsmentioning
confidence: 99%
See 2 more Smart Citations