2018 International Conference on Advanced Technologies for Communications (ATC) 2018
DOI: 10.1109/atc.2018.8587580
|View full text |Cite
|
Sign up to set email alerts
|

Fixed-Point Implementation of Convolutional Neural Networks for Image Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 26 publications
(11 citation statements)
references
References 1 publication
0
9
0
Order By: Relevance
“…The extent of these complications has introduced shallow networks [525], approaches to quantizing model parameters [526] along with ternary [527] and binary [528] models that focus on reducing the memory overhead for efficient resource-constrained hardware accelerator implementation. Authors in [529] provide an example of a fixed-point CNN classifier involving 4 bit fixed point arithmetic that suggests negligible accuracy degradation and authors in [530] present fast BNN inference accelerators to meet the FPGA on-chip memory requirements. Reducing memory footprints in hardware accelerators is also tied up to the cost-effective designing of memory units.…”
Section: Current and Future Challengesmentioning
confidence: 99%
“…The extent of these complications has introduced shallow networks [525], approaches to quantizing model parameters [526] along with ternary [527] and binary [528] models that focus on reducing the memory overhead for efficient resource-constrained hardware accelerator implementation. Authors in [529] provide an example of a fixed-point CNN classifier involving 4 bit fixed point arithmetic that suggests negligible accuracy degradation and authors in [530] present fast BNN inference accelerators to meet the FPGA on-chip memory requirements. Reducing memory footprints in hardware accelerators is also tied up to the cost-effective designing of memory units.…”
Section: Current and Future Challengesmentioning
confidence: 99%
“…Deep learning algorithms are usually implemented in software using 32 bit floating-point values (FP32). Migrating a deep learning algorithm to an ASIC or an FPGA requires for a bit width reduction which is possible using the quantization technique [ 24 , 25 , 26 ]. A quantized model and a non-quantized model execute the same operations, however, a quantized model with bit-width reduction promotes a memory reduction and allows the execution of more operations per cycle.…”
Section: State Of the Artmentioning
confidence: 99%
“…For instance, an 8-bit integer multiplier is much faster than its 32-bit floating-point counterpart. Some existing quantization techniques require retraining of the network (LQ-Nets [7], [15]), which is not flexible in many occasions. Hence, this work only focuses on post-training static quantization that does not require retraining.…”
Section: Background and Motivation Of Doubleq A Quantization Of Convolutional Neural Network (Cnn)mentioning
confidence: 99%