2022
DOI: 10.1007/s10489-022-03348-z
|View full text |Cite
|
Sign up to set email alerts
|

Self-distribution binary neural networks

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(6 citation statements)
references
References 36 publications
0
6
0
Order By: Relevance
“…Another binary network based on MobileNetV1 is ReActNet [87] which proposed ReAct-Sign (RSign) and ReAct-PReLU (RPReLU) as alternatives of the traditional activation function to reshape the activation distribution. While in [88], the authors proposed Activation Self Distribution (ASD) and Weight Self Distribution (WSD) to adjust the sign distribution of activations and weights, respectively, to enhance the accuracy.…”
Section: C: Gradient Error Minimizationmentioning
confidence: 99%
“…Another binary network based on MobileNetV1 is ReActNet [87] which proposed ReAct-Sign (RSign) and ReAct-PReLU (RPReLU) as alternatives of the traditional activation function to reshape the activation distribution. While in [88], the authors proposed Activation Self Distribution (ASD) and Weight Self Distribution (WSD) to adjust the sign distribution of activations and weights, respectively, to enhance the accuracy.…”
Section: C: Gradient Error Minimizationmentioning
confidence: 99%
“…We first evaluate our method on CIFAR10 dataset by comparing it with existing state-of-the-arts binary quantization methods, including DSQ [7], DoReFa [22], IR-Net [16], L2B [21] and SD-BNN [20]. Among them, DSQ [7] and IR-Net [16] propose to use T anh as soft functions to approximate the Sign function.…”
Section: Cifar10mentioning
confidence: 99%
“…We will further verify this view in section 5.3. 1/1 85.3 IR-Net [16] 1/1 86.5 SD-BNN [20] 1/1 86.9 Ours 1/1 87.9…”
Section: Cifar10mentioning
confidence: 99%
“…This implies the new weight distribution has an equal amount of weights below and above zero, and therefore after binarization, the amount of -1's and +1's is equal and entropy is maximized. SD-BNN (2021) [50] employs a series of linear layers and non-linearities to calculate biases for both weights and features. Note that this approach for the features belongs in the category of network topology changing.…”
Section: Normalizationmentioning
confidence: 99%
“…R20C10 experiments will use default data augmentation and train for 500 epochs, whereas R18C100 experiments will use AutoAug data augmentation and train for 1000 epochs. [46] 91.10% Circulant BNN [28] 69.97% IR-Net [34] 86.5% Real-to-Binary [32] 76.2% BBG [39] 85.34% ProxyBNN [17] 67.17% RBNN [25] 87.8% Information capacity BNN [19] 73.48% Noisy Supervision [14] 85.78% ReCU [49] 69.1% ReCU [49] 87.4% BNN-BN [8] 68.34% Sub-bit BNN [45] 83.9% Equal Bits [24] 71.60% SD-BNN [50] 86.9% BNN fully latent weights [48] 88.6% Fig. 16 Baseline BNN compared to different feature binarizers.…”
Section: Design Space Explorationmentioning
confidence: 99%