The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2015
DOI: 10.48550/arxiv.1511.00363
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

BinaryConnect: Training Deep Neural Networks with binary weights during propagations

Abstract: Deep Neural Networks (DNN) have achieved state-of-the-art results in a wide range of tasks, with the best results obtained with large training sets and large models. In the past, GPUs enabled these breakthroughs because of their greater computational speed. In the future, faster computation at both training and test time is likely to be crucial for further progress and for consumer applications on low-power devices. As a result, there is much interest in research and development of dedicated hardware for Deep … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
92
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 53 publications
(92 citation statements)
references
References 20 publications
0
92
0
Order By: Relevance
“…Model compression has been extensively studied especially for image-classification tasks, see e.g., [38], [39], [40], [41], [42]. The typical model compression techniques include weight quantization/sparsification [38], [41], [42], network pruning [43], [44], KD [40], [45], and lightweight neural network architecture/operation design [46], [47], [48]. In this work, we mainly focus on model compression for vanilla GANs, i.e., noise-to-image task.…”
Section: Related Workmentioning
confidence: 99%
“…Model compression has been extensively studied especially for image-classification tasks, see e.g., [38], [39], [40], [41], [42]. The typical model compression techniques include weight quantization/sparsification [38], [41], [42], network pruning [43], [44], KD [40], [45], and lightweight neural network architecture/operation design [46], [47], [48]. In this work, we mainly focus on model compression for vanilla GANs, i.e., noise-to-image task.…”
Section: Related Workmentioning
confidence: 99%
“…The concept of Binary Neural Network (BNN) originated from the binary weight neural network (BWNN) [18], and the BWNN only quantizes the bit representation of the weight value into the binary value. However, for the FPGA devices with small on-chip memory, the intermediate activations of the BWNN are still too large to be stored in the on-chip SRAM, and external memory is required.…”
Section: B Binary Complex Neural Networkmentioning
confidence: 99%
“…4) Binarization: There are two types of widely used binarization [18]: deterministic binarization and stochastic binarization. The equation for deterministic binarization is given in Eq.…”
Section: B Building Blocks and Operationsmentioning
confidence: 99%
See 1 more Smart Citation
“…Quantization-aware training, which directly trains the network with lower precisions [6]. These approaches progressively enabled DNNs to first be quantized to 16-bit fixed point [7], 8-bit fixed point [8], and all the way down to binary precision [9]. The best precision of DNN parameters, however, varies across different NN models, and even across different layers within one model [5], [10].…”
Section: Introductionmentioning
confidence: 99%