2018 IEEE Computer Society Annual Symposium on VLSI (ISVLSI) 2018
DOI: 10.1109/isvlsi.2018.00034
|View full text |Cite
|
Sign up to set email alerts
|

TaiJiNet: Towards Partial Binarized Convolutional Neural Network for Embedded Systems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 3 publications
0
5
0
Order By: Relevance
“…Low-rank decomposition [22], [23] Reduce the matrix size Effectively reduce storage and consumption,parameter is related to the number of network layers and is not effective for large-scale networks Pruning [24]- [26] Remove unimportant parameters High precision, less parameters, prevent over fitting,time-consuming and computational and the network is unstructured Quantization Convert floating point arithmetic to fixed point operation More concise model, difficult to implement , instability of accuracy, poor versatility Knowledge distillation [27], [28] [28] Transferring learning, training a small network Suitable for small model training ,artificial design seriously affects the training effect Binaryzation [12] Parameter binarization Greatly reduced storage,simple calculation, poor model performance function of the Jth channel of the output is shown in formula (1).…”
Section: Methods Description Advantages and Disadvantagesmentioning
confidence: 99%
See 2 more Smart Citations
“…Low-rank decomposition [22], [23] Reduce the matrix size Effectively reduce storage and consumption,parameter is related to the number of network layers and is not effective for large-scale networks Pruning [24]- [26] Remove unimportant parameters High precision, less parameters, prevent over fitting,time-consuming and computational and the network is unstructured Quantization Convert floating point arithmetic to fixed point operation More concise model, difficult to implement , instability of accuracy, poor versatility Knowledge distillation [27], [28] [28] Transferring learning, training a small network Suitable for small model training ,artificial design seriously affects the training effect Binaryzation [12] Parameter binarization Greatly reduced storage,simple calculation, poor model performance function of the Jth channel of the output is shown in formula (1).…”
Section: Methods Description Advantages and Disadvantagesmentioning
confidence: 99%
“…Based on network compression, an algorithm compares the effects of precision and speed on different resolution weights, and it is shown that the lower the number of bits of the weight is, the lower the accuracy obtained. The network compresses the weight or activation value to one bit, and the compression ratio is 32:1 [10]- [12], [20]. [12] replaced some layers with the binary network layer.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Most recently, hybrid quantization has attracted more and more attention, because it enables better trade-off between compression and performance [36]- [38]. As for partial binarization, a sub-area of hybrid quantization, on which we are focused, both training methods [39] and the corresponding hardware accelerators [19], [40] are also investigated extensively. The actual performance after compression heavily depends on the configuration of the partial binarization, i.e.…”
Section: A Cnn Compressionmentioning
confidence: 99%
“…Naive quantization usually leads to total failure especially for binarization. Significant effort has been devoted to develop better quantization and binarization methods as well as the hardware accelerator [14], [16]- [19]. Its success on CNNs has been demonstrated by multiple works, where memory consumption is deeply compressed although sometimes the performance cannot be preserved [20]- [23].…”
Section: Introductionmentioning
confidence: 99%