2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00732
|View full text |Cite
|
Sign up to set email alerts
|

A Main/Subsidiary Network Framework for Simplifying Binary Neural Networks

Abstract: To reduce memory footprint and run-time latency, techniques such as neural network pruning and binarization have been explored separately. However, it is unclear how to combine the best of the two worlds to get extremely small and efficient models. In this paper, we, for the first time, define the filter-level pruning problem for binary neural networks, which cannot be solved by simply migrating existing structural pruning methods for full-precision models. A novel learning-based approach is proposed to prune … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0
3

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 26 publications
(15 citation statements)
references
References 18 publications
0
12
0
3
Order By: Relevance
“…The process of knowledge distillation is shown in Figure 3. Similar mimic solutions like Distillation and Quantization (DQ) [81], Distilled Binary Neural Network (DBNN) [80] and Main/Subsidiary Network [87] have been studied, and their experiments demonstrate that the loss functions related to the full-precision teacher model help to stabilize the training of binary student model with high accuracy. CI-BCNN proposed in [86] mines the channel-wise interactions, through which prior knowledge is provided to alleviate inconsistency of signs in binary feature maps and preserves the information of input samples during inference.…”
Section: Improve the Network Loss Functionmentioning
confidence: 99%
“…The process of knowledge distillation is shown in Figure 3. Similar mimic solutions like Distillation and Quantization (DQ) [81], Distilled Binary Neural Network (DBNN) [80] and Main/Subsidiary Network [87] have been studied, and their experiments demonstrate that the loss functions related to the full-precision teacher model help to stabilize the training of binary student model with high accuracy. CI-BCNN proposed in [86] mines the channel-wise interactions, through which prior knowledge is provided to alleviate inconsistency of signs in binary feature maps and preserves the information of input samples during inference.…”
Section: Improve the Network Loss Functionmentioning
confidence: 99%
“…BinaryConnect [17] 1/32 VGG-Small 91.7 BNN [18] 1/1 VGG-Small 89.9 XNOR-Net [19] 1/1 VGG-Small 89.8 LQ-Nets [20] 1/32 ResNet-20 90.1 BBG [21] 1/1 ResNet-20 85.3 BCGD [22] 1/4 VGG-11 89.6 IR-Net [23] 1/1 VGG-Small 90.4 CI-BCNN [24] 1/1 VGG-Small 92.5 Multi-scale BNN(Ours)…”
Section: Methodsmentioning
confidence: 99%
“…6, the binarized CompConv beat binarized Ghost both in parameters and FLOPs, with nearly 2× acceleration and storage saving. Though the binarized model is lightweight enough, there still exists redundancy of computation cost as pointed by [38], the computation efficiency of which can also be improved by replacing regular convolution with our CompConv. We also apply CompConv on ResNet to verify its capability of finer classification on CIFAR-100.…”
Section: Binarized Compconv On Cifar-10mentioning
confidence: 99%