2022
DOI: 10.1088/2634-4386/ac7c8a
|View full text |Cite
|
Sign up to set email alerts
|

Two sparsities are better than one: unlocking the performance benefits of sparse–sparse networks

Abstract: In principle, sparse neural networks should be significantly more efficient than traditional dense networks. Neurons in the brain exhibit two types of sparsity; they are sparsely interconnected and sparsely active. These two types of sparsity, called weight sparsity and activation sparsity, when combined, offer the potential to reduce the computational cost of neural networks by two orders of magnitude. Despite this potential, today’s neural networks deliver only modest performance benefits using just weight s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 70 publications
(96 reference statements)
0
4
0
Order By: Relevance
“…By contrast, using DN on RNNs is beneficial because the fully-connected RNNs are weight-memory bounded, and the energy saving brought by temporal sparsity is much larger for RNNs. [7] and [10] exploit both DN activation sparsity and weight sparsity to achieve impressive inference performance on hardware. Another method that can create sparsity in neural networks is conditional computation, or skipping operations.…”
Section: Related Workmentioning
confidence: 99%
“…By contrast, using DN on RNNs is beneficial because the fully-connected RNNs are weight-memory bounded, and the energy saving brought by temporal sparsity is much larger for RNNs. [7] and [10] exploit both DN activation sparsity and weight sparsity to achieve impressive inference performance on hardware. Another method that can create sparsity in neural networks is conditional computation, or skipping operations.…”
Section: Related Workmentioning
confidence: 99%
“…Activity sparsity in RNNs has been proposed previously in various forms [29,47,48], but only focusing on achieving it during inference. Conditional computation is a form of activity sparsity used in [17] to scale to 1 trillion parameters.…”
Section: Related Workmentioning
confidence: 99%
“…Further, Moraitis et al (2022) introduce an unsupervised local training algorithm based on a combination of Hebbian plasticity with a soft winner take all mechanism. On the hardware side, two studies introduce new methods to exploit sparsity on current hardware (graphics processing unit (GPU) and field programmable gate array (FPGA)) to improve inference efficiency through Complementary Sparsity (Turner et al 2022) and Procedural connectivity (Hunter et al 2022). Finally, on the application side, DeWolf et al (2023) introduce a welcome closed-loop benchmark control task based on a robotic arm simulated in the popular Mojoco platform to showcase the inherent power efficiency and low latency of event-based computation.…”
mentioning
confidence: 99%
“…Although the sparse activity and connectivity of SNNs should result in a proportional reduction in computing requirements, the irregular patterns of neuron interconnectivity and activity reduce the expected gains on current hardware. Hunter et al (2022) address this problem by structuring the sparsity to match the requirements of the target hardware for implementing sparse activation-sparse connectivity networks on FPGA. This restructuring is achieved by overlaying multiple sparse matrices to form a single dense structure if no two sparse matrices contain non-zero elements at the same location.…”
mentioning
confidence: 99%