2021
DOI: 10.48550/arxiv.2112.00133
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

PokeBNN: A Binary Pursuit of Lightweight Accuracy

Abstract: Top-1 ImageNet optimization promotes enormous networks that may be impractical in inference settings. Binary neural networks (BNNs) have the potential to significantly lower the compute intensity but existing models suffer from low quality. To overcome this deficiency, we propose Poke-Conv, a binary convolution block which improves quality of BNNs by techniques such as adding multiple residual paths, and tuning the activation function. We apply it to ResNet-50 and optimize ResNet's initial convolutional layer … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 33 publications
(51 reference statements)
0
3
0
Order By: Relevance
“…The research in BNNs is focussed on bringing deep learning to resourceconstrained edge devices. Recent studies report the computational complexity of their models using theoretical metrics such as floating-point operations (FLOPs) [24,27] multiplyaccumulate (MACs) [4] or arithmetic computation effort (ACE) [40]. In coherence with [3,30] we argue that latency is the best metric to compare model performances.…”
Section: Methodsmentioning
confidence: 99%
“…The research in BNNs is focussed on bringing deep learning to resourceconstrained edge devices. Recent studies report the computational complexity of their models using theoretical metrics such as floating-point operations (FLOPs) [24,27] multiplyaccumulate (MACs) [4] or arithmetic computation effort (ACE) [40]. In coherence with [3,30] we argue that latency is the best metric to compare model performances.…”
Section: Methodsmentioning
confidence: 99%
“…Note that while the compute cost of a SE-like module is usually considered to be negligible, its parameter size cannot be ignored in BNN models. We calculated total number of opertations (OPs) as OPs = FLOPs + (BOPs / 64) + (int4 OPs / 16), following [18,19,32]. In case of parameters, binary weights are 1-bit, weights of SE-like modules are 8-bit, and other real-valued parameters and weights are considered as 32-bit.…”
Section: Cost Analysismentioning
confidence: 99%
“…We use Arithmetic Computation Effort (ACE) (Zhang et al, 2021), a newly proposed hardware-and energy-inspired cost metric to evaluate the inference cost of quantized BERT models. ACE is defined as Table 1: We quantize 32-bit baseline models to 8-bits by three quantization methods.…”
Section: Downstream Language Tasks and Evaluation Metricsmentioning
confidence: 99%
“…I and J are sets of all quantization bits used for inference. ACE is shown to be well correlated to the actual energy consumption on Google TPUs hardware and used to evaluate the inference cost of Binary Neural Networks in Zhang et al (2021).…”
Section: Downstream Language Tasks and Evaluation Metricsmentioning
confidence: 99%