2018
DOI: 10.48550/arxiv.1812.11800
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Regularized Binary Network Training

Abstract: The deployment of Deep neural networks (DNN) on edge devices has been difficult because they are resource hungry. Binary neural networks (BNN) help to alleviate the prohibitive resource requirements of DNN, where both activations and weights are limited to 1-bit. There is however a significant performance gap between BNNs and floating point DNNs. To reduce this gap, We propose an improved binary training method, by introducing a new regularization function that encourages training weights around binary values.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
50
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 22 publications
(51 citation statements)
references
References 6 publications
0
50
0
Order By: Relevance
“…But our proposed RA-BNN outperforms both 8-bit and binary baseline, as even after 5000 bit flips the accuracy only degrades to 37.1 % on ImageNet. [22] 53.0 72.6 RBNN [19] 59.9 81.9 Bi-Real [21] 56.4 79.5 RA-BNN 62.9 84.1…”
Section: Robustness Evaluationmentioning
confidence: 99%
See 2 more Smart Citations
“…But our proposed RA-BNN outperforms both 8-bit and binary baseline, as even after 5000 bit flips the accuracy only degrades to 37.1 % on ImageNet. [22] 53.0 72.6 RBNN [19] 59.9 81.9 Bi-Real [21] 56.4 79.5 RA-BNN 62.9 84.1…”
Section: Robustness Evaluationmentioning
confidence: 99%
“…However, for a binary weight neural network, with a sufficiently large amount of attack iterations, the attacker can still successfully degrade its accuracy to as low as random guess [16]. More importantly, due to aggressively compressing the floating-point weights (i.e., 32 bits or more) into binary (1 bit), BNN inevitably sacrifices its clean model accuracy by 10-30 %, which is widely discussed in prior works [19][20][21][22][23]. Therefore, in this work, for the first time, our primary goal is to construct a robust and accurate binary neural network (with both binary weight and activation) to simultaneously defend bit-flip attacks and improve clean model accuracy.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Obviously, the memory usage for weights and activations Figure 1: From top to bottom: original functions in spatial domain, corresponding functions in frequency domain and the difference between the current function and sign function in frequency domain. From left to right: sign function, combination of sine functions, tanh function in [8] and SignSwish function in [5] (short as SS).…”
Section: Introductionmentioning
confidence: 99%
“…For example, DSQ [8] introduced a tanh-alike differentiable asymptotic function to estimate the forward and backward procedures of the conventional sign function. BNN+ [5] used a SignSwish activation function to modify the back-propagation of the original sign function and further introduced a regularization that encourages the weights around binary values. RBNN [22] proposed a training-aware approximation function to replace the sign function when computing the gradient.…”
Section: Introductionmentioning
confidence: 99%