2021
DOI: 10.21203/rs.3.rs-133395/v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Truly Sparse Neural Networks at Scale

Abstract: Recently, sparse training methods have started to be established as a de facto approach for training and inference efficiency in artificial neural networks. Yet, this efficiency is just in theory. In practice, everyone uses a binary mask to simulate sparsity since the typical deep learning software and hardware are optimized for dense matrix operations. In this paper, we take an orthogonal approach, and we show that we can train truly sparse neural networks to harvest their full potential. To achieve this goal… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(1 citation statement)
references
References 35 publications
0
1
0
Order By: Relevance
“…Sparse neuron architectures can be achieved by other means: Mollaysa et al [2017] enforce sparsity based on the Jacobian and Li et al [2016], Lee et al [2006], Ranzato et al [2007], Collins and Kohli [2014], Ma et al [2019] employ 1 -based LASSO penalty to induce sparsity. Curci et al [2021] prune their ANNs based on a metric for neuron importance. Evci et al [2019] discuss the difficulty of training sparse ANNs.…”
Section: Introductionmentioning
confidence: 99%
“…Sparse neuron architectures can be achieved by other means: Mollaysa et al [2017] enforce sparsity based on the Jacobian and Li et al [2016], Lee et al [2006], Ranzato et al [2007], Collins and Kohli [2014], Ma et al [2019] employ 1 -based LASSO penalty to induce sparsity. Curci et al [2021] prune their ANNs based on a metric for neuron importance. Evci et al [2019] discuss the difficulty of training sparse ANNs.…”
Section: Introductionmentioning
confidence: 99%