2022
DOI: 10.1145/3510026
|View full text |Cite
|
Sign up to set email alerts
|

Block Walsh–Hadamard Transform-based Binary Layers in Deep Neural Networks

Abstract: Convolution has been the core operation of modern deep neural networks. It is well-known that convolutions can be implemented in the Fourier Transform domain. In this paper, we propose to use binary block Walsh-Hadamard transform (WHT) instead of the Fourier transform. We use WHT-based binary layers to replace some of the regular convolution layers in deep neural networks. We utilize both one-dimensional (1-D) and two-dimensional (2-D) binary WHTs in this paper. In both 1-D and 2-D layers, we compute the binar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

2
4

Authors

Journals

citations
Cited by 9 publications
(3 citation statements)
references
References 42 publications
0
3
0
Order By: Relevance
“…WHT presents a computational challenge when the dimension of the input vector is not a power of two. A technique called blockwise WHTs (BWHTs) was introduced to address this issue in [28]. The BWHT approach divides the transform matrix into multiple blocks, each sized to an integer power of two.…”
Section: Background a Walsh-hadamard Transformmentioning
confidence: 99%
See 2 more Smart Citations
“…WHT presents a computational challenge when the dimension of the input vector is not a power of two. A technique called blockwise WHTs (BWHTs) was introduced to address this issue in [28]. The BWHT approach divides the transform matrix into multiple blocks, each sized to an integer power of two.…”
Section: Background a Walsh-hadamard Transformmentioning
confidence: 99%
“…On the other hand, the projection operation employs a 1-D-BWHT layer to reduce the dimensional to make the network computationally efficient while retaining essential features. In Pan et al's study [28], these transformations maintained a matching accuracy under frequency transforms while achieving significant compression than standard implementation on benchmark datasets such as CIFAR-10, CIFAR-100, and ImageNet. The number of parameters in the BWHT layer is thus proportional to the thresholding parameter T , which is significantly smaller than the number of parameters in a 1 × 1 convolution layer.…”
Section: B Frequency-domain Compression Of Deep Neural Networkmentioning
confidence: 99%
See 1 more Smart Citation