Proceedings of the 54th Annual Design Automation Conference 2017 2017
DOI: 10.1145/3061639.3062189
|View full text |Cite
|
Sign up to set email alerts
|

A Kernel Decomposition Architecture for Binary-weight Convolutional Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
11
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 31 publications
(11 citation statements)
references
References 7 publications
0
11
0
Order By: Relevance
“…FINN [53] uses FPGAs for accelerating Binary DNNs, while YodaNN [54] and BRein [55] propose an ASIC accelerator for binary DNNs. Kim, et al [56] decompose the convolution weights for binary CNNs to improve performance and energy efficiency. The above works focus solely on binary DNNs to achieve high performance at the cost of classification accuracy.…”
Section: Related Workmentioning
confidence: 99%
“…FINN [53] uses FPGAs for accelerating Binary DNNs, while YodaNN [54] and BRein [55] propose an ASIC accelerator for binary DNNs. Kim, et al [56] decompose the convolution weights for binary CNNs to improve performance and energy efficiency. The above works focus solely on binary DNNs to achieve high performance at the cost of classification accuracy.…”
Section: Related Workmentioning
confidence: 99%
“…[17] investigated the opportunity to use deep learning for identifying nonintuitive features from cross-sensor correlations by means of an RBM. In [1] a kernel decomposing scheme in binary-weight networks is proposed that skips redundant computations and achieves 22% energy reduction on image classification. BiNMAC is proposed in [18] which is a programmable manycore accelerator for BNNs designed for physiological and Image processing case studies.…”
Section: Related Workmentioning
confidence: 99%
“…Such devices usually process real-time data, read from multimodal sensors continuously, and suffer from resource-bound and limited battery budget due to their small size, online monitoring, and portability. Therefore, minimizing the power dissipation of these devices while meeting real-time requirements is a subject of interest [1][2][3][4][5].…”
Section: Introductionmentioning
confidence: 99%
“…As a distinctive feature, the binary quantization is not only applied during the forward pass, but also during the backward pass of the gradient descent algorithm, and acts as a sort of regularizer [16]. Hardware accelerators for highly-quantized NNs have been presented on FPGA [23], ASIC [3,10] and neuromorphic brain-inspired chips such as Truenorth [6], trading the exibility of general-purpose processors with highest performance and energy e ciency of specialized hardware. To lower the computational complexity of BNNs, a hardware-oriented kernel decomposition strategy is presented in [10], using clockgating to reduce the energy cost of redundant convolutions.…”
Section: Related Workmentioning
confidence: 99%
“…Hardware accelerators for highly-quantized NNs have been presented on FPGA [23], ASIC [3,10] and neuromorphic brain-inspired chips such as Truenorth [6], trading the exibility of general-purpose processors with highest performance and energy e ciency of specialized hardware. To lower the computational complexity of BNNs, a hardware-oriented kernel decomposition strategy is presented in [10], using clockgating to reduce the energy cost of redundant convolutions. is is clearly e ective but less prone to be implemented in so ware because it is weight-dependent and does not bene t from data spatial contiguity, which is exploited in this work to reduce the computation latency.…”
Section: Related Workmentioning
confidence: 99%