The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2021
DOI: 10.48550/arxiv.2103.00841
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning Frequency Domain Approximation for Binary Neural Networks

Abstract: Binary neural networks (BNNs) represent original full-precision weights and activations into 1-bit with sign function. Since the gradient of the conventional sign function is almost zero everywhere which cannot be used for back-propagation, several attempts have been proposed to alleviate the optimization difficulty by using approximate gradient. However, those approximations corrupt the main direction of de facto gradient. To this end, we propose to estimate the gradient of sign function in the Fourier freque… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 23 publications
0
2
0
Order By: Relevance
“…SI-BNN [79] proposed trainable parameters for activations and gradients in the back-propagation. In addition, the Fourier Frequency Domain Approximation (FDA) is used to update the gradients in back-propagation in [80]. J. Lee et al [81] introduced an Element-Wise Gradient Scaling (EWGS) to update each gradient element by scaling factor.…”
Section: C: Gradient Error Minimizationmentioning
confidence: 99%
“…SI-BNN [79] proposed trainable parameters for activations and gradients in the back-propagation. In addition, the Fourier Frequency Domain Approximation (FDA) is used to update the gradients in back-propagation in [80]. J. Lee et al [81] introduced an Element-Wise Gradient Scaling (EWGS) to update each gradient element by scaling factor.…”
Section: C: Gradient Error Minimizationmentioning
confidence: 99%
“…General model compression approaches fall under multiple forms [12]: pruning [21,63], quantization [64,56,16], knowledge distillation [27,44], as well as their compositions [61,69,71]. A Binary Neural Network (BNN) [13,14,34,73,51,14,36,48,44,29,72,39,25,37,10,57,20,66] represents the most extreme form of model quantization as it quantizes weights in convolution layers to only 1 bit, enjoying great speed-up compared with its full-precision counterpart. [50] roughly divides previous BNN literature into two categories: (i) native BNN [13,14,34] which directly applies binarization to a full-precision model by a pre-defined binarization function.…”
Section: Related Workmentioning
confidence: 99%