2018 IEEE International Symposium on High Performance Computer Architecture (HPCA) 2018
DOI: 10.1109/hpca.2018.00016
|View full text |Cite
|
Sign up to set email alerts
|

Towards Efficient Microarchitectural Design for Accelerating Unsupervised GAN-Based Deep Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
24
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 51 publications
(25 citation statements)
references
References 17 publications
0
24
0
Order By: Relevance
“…We observe that stereo DNNs make heavy use of the deconvolution operation 1 that exposes specific kernel sparsity, making conventional DNN accelerators inefficient. While prior work proposed specialized hardware to exploit deconvolution sparsity [60,76], we demonstrate that static software optimizations achieve better results without unnecessary hardware modifications.…”
Section: Introductionmentioning
confidence: 81%
See 1 more Smart Citation
“…We observe that stereo DNNs make heavy use of the deconvolution operation 1 that exposes specific kernel sparsity, making conventional DNN accelerators inefficient. While prior work proposed specialized hardware to exploit deconvolution sparsity [60,76], we demonstrate that static software optimizations achieve better results without unnecessary hardware modifications.…”
Section: Introductionmentioning
confidence: 81%
“…Stereo vision DNNs make use of deconvolution layers, which expose structured sparsity patterns. Recent work has prosed specialized hardware specifically for exploiting sparsity in deconvolution layers [60,76]. Our observation, however, is that mitigating sparsityinduced efficiencies in deconvolution does not necessarily require hardware support.…”
Section: Related Workmentioning
confidence: 98%
“…Deconvolution is an operation that is also adopted in generative adversarial networks (GAN), and there have been many studies on deconvolution accelerators for GANs. Wang et al [37], Song et al [38], GANAX [39], LerGAN [40] accelerated the 16-bit network, and GNA [41] implemented deconvolution accelerator that supports flexible bit-width of 8-bit and 16-bit.…”
Section: Related Workmentioning
confidence: 99%
“…References [22]- [24] consider the padding-zero operations when designing an accelerator, but they do not eliminate the operations. Reference [25] reshapes the input data to skip the zero-data, but it is customized for the GAN used in its work; it may need the filling-zero operations when it computes conventional CNNs with different kernel sizes.…”
Section: A Padding-zero Operationsmentioning
confidence: 99%