2018
DOI: 10.1007/978-3-030-01261-8_39
|View full text |Cite
|
Sign up to set email alerts
|

Training Binary Weight Networks via Semi-Binary Decomposition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(5 citation statements)
references
References 18 publications
0
5
0
Order By: Relevance
“…However, the SOTA neural network models suffer massive parameters and large sizes to achieve good performance in different tasks, which also cause significant complex computation and great resource consumption. To compress and accelerate the deep CNNs, many approaches have been proposed, which can be classified into five categories: transferred/compact convolutional filters [89,85,78]; quantization/binarization [35,11,82,92]; knowledge distillation [12,86,16]; pruning [28,31,22]; low-rank factorization [46,38,47,79].…”
Section: Related Workmentioning
confidence: 99%
“…However, the SOTA neural network models suffer massive parameters and large sizes to achieve good performance in different tasks, which also cause significant complex computation and great resource consumption. To compress and accelerate the deep CNNs, many approaches have been proposed, which can be classified into five categories: transferred/compact convolutional filters [89,85,78]; quantization/binarization [35,11,82,92]; knowledge distillation [12,86,16]; pruning [28,31,22]; low-rank factorization [46,38,47,79].…”
Section: Related Workmentioning
confidence: 99%
“…parameter quantizing [32,33,34,35,36,37,38,39,40,41], low-rank parameter factorization [42,43,44,45,46], transferred/compact convolutional filters [47,48,49,50], and knowledge distillation [51,52,53,54,55,56]. The parameter pruning and quantizing mainly focus on eliminating the redundancy in the model parameters respectively by removing the redundant/uncritical ones or compressing the parameter space (e.g.…”
Section: Introductionmentioning
confidence: 99%
“…They visualize the feature maps extracted by different filters and view each filter as a visual unit focusing on different visual components.of the ResNet-50 [28], and meanwhile save more than 75% of parameters and 50% computational time. In the literature, approaches for compressing the deep networks can be classified into five categories: parameter pruning [26,29,30,31], parameter quantizing [32,33,34,35,36,37,38,39,40,41], low-rank parameter factorization [42,43,44,45,46], transferred/compact convolutional filters [47,48,49,50], and knowledge distillation [51,52,53,54,55,56]. The parameter pruning and quantizing mainly focus on eliminating the redundancy in the model parameters respectively by removing the redundant/uncritical ones or compressing the parameter space (e.g.…”
mentioning
confidence: 99%
“…By using FP16, the final accuracy is determined by the selected network and the corresponding BWN training algorithm. A ResNet-18 trained on the ImageNet dataset can run on Hyperdrive with a 87.1% top-5 accuracy using the SBD-FQ training method [55] (full-precision top-5 accuracy: 89.2%).…”
Section: Resultsmentioning
confidence: 99%