2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2021
DOI: 10.1109/cvprw53098.2021.00140
|View full text |Cite
|
Sign up to set email alerts
|

A Bop and Beyond: A Second Order Optimizer for Binarized Neural Networks

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 8 publications
0
3
0
Order By: Relevance
“…Yang (2020) takes this one step further and introduces the Filter Gradient descent Framework that can use different types of filters on the noisy gradients to make a better estimation of the true gradient. In binary network optimization, Bop (Helwegen et al, 2019) and its extension (Suarez-Ramirez et al, 2021) introduce a threshold to compare with the smoothed gradient by EMA to determine whether to flip a binary weight. In our paper, we build on second order gradient filtering techniques to reinterpret the hyperparameters that influence the latent weight updates.…”
Section: Related Workmentioning
confidence: 99%
“…Yang (2020) takes this one step further and introduces the Filter Gradient descent Framework that can use different types of filters on the noisy gradients to make a better estimation of the true gradient. In binary network optimization, Bop (Helwegen et al, 2019) and its extension (Suarez-Ramirez et al, 2021) introduce a threshold to compare with the smoothed gradient by EMA to determine whether to flip a binary weight. In our paper, we build on second order gradient filtering techniques to reinterpret the hyperparameters that influence the latent weight updates.…”
Section: Related Workmentioning
confidence: 99%
“…Obviously, this introduces a gradient mismatch between the actual gradient of the [18] 2019 n/a Custom BENN (6x) [55] 2019 61 LC_1 LC_1 N ADAM EN n/a N N Circulant BNN [28] 2019 61.4 PN BN PN N SGD AM PReLU Y N CI-BCNN [46] 2019 59.9 [19] 2020 61.36 [44] 2020 59.7 [29] 2021 n/a PN LB LC_1 TO Y ADAM AM PReLU Y N ReCU [49] 2021 66.4 PN STD LC_A MSTDB N SGD LF PReLU Y N Bop and beyond [41] 2021 n/a Custom Sub-bit BNN [45] 2021 55 In recent years, there have been works that change the clipping interval of the STE. For example, BinaryDenseNet [4] and MeliusNet [3] use an interval of [−1.3, +1.3], whereas PokeBNN [53] uses an interval of [−3, +3].…”
Section: Binarizer (Ste)mentioning
confidence: 99%
“…Next to the default optimizers, [18,41] have developed new optimizers that are dedicated to BNNs, which are respectively called Bop and Bop2ndOrder.…”
Section: Optimizermentioning
confidence: 99%