2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2016
DOI: 10.1109/cvpr.2016.521
|View full text |Cite
|
Sign up to set email alerts
|

Quantized Convolutional Neural Networks for Mobile Devices

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

0
574
0
6

Year Published

2017
2017
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 980 publications
(580 citation statements)
references
References 19 publications
0
574
0
6
Order By: Relevance
“…To tackle the storage issue of deep networks, Gong et al [6], Wu et al [31] and Lin et al [18] consider applying the quantization techniques to pre-trained CNNs, and trying to make network compressions with minor concessions on the inference accuracy. Another powerful category of methods in this scope is network pruning.…”
mentioning
confidence: 99%
“…To tackle the storage issue of deep networks, Gong et al [6], Wu et al [31] and Lin et al [18] consider applying the quantization techniques to pre-trained CNNs, and trying to make network compressions with minor concessions on the inference accuracy. Another powerful category of methods in this scope is network pruning.…”
mentioning
confidence: 99%
“…Therefore, we introduce our approach on adapting SqueezeNet (Iandola et al, 2016), a smaller CNN with a model size of only 4,8 MB which is even 10 times smaller than GoogLeNet. In future work this might be reduced even more by quantizing (Wu et al, 2016) or binarization (Courbariaux et al, 2016).…”
Section: Related Workmentioning
confidence: 99%
“…Such techniques focus on compressing an already trained CNN, instead of training a CNN with fewer parameters in the first place. Some of these works [7], [9], also use vector quantization techniques. However, proposed the method uses a differentiable quantization scheme that allows for training both the quantizer and the rest of the network simultaneously.…”
Section: Related Workmentioning
confidence: 99%
“…Many techniques have been proposed to reduce the model size [6]- [9]. Usually compression and pruning techniques are used to reduce the size of CNN models [6], [7], [9]. Such techniques focus on compressing an already trained CNN, instead of training a CNN with fewer parameters in the first place.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation