2020
DOI: 10.1016/j.neucom.2020.02.035
|View full text |Cite
|
Sign up to set email alerts
|

Sparse low rank factorization for deep neural network compression

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
46
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 96 publications
(69 citation statements)
references
References 14 publications
0
46
0
Order By: Relevance
“…Low-rank decomposition: Low-rank decomposition algorithms [ 30 , 31 , 32 ] use a lower-rank set instead of the original set of parameters to approximate the CNN to achieve compression. Swaminathan et al [ 31 ] argue that the low-rank decomposition of weight matrices should consider influence of both input as well as output neurons of a layer. They propose a sparse low rank (SLR) approach that sparsifies SVD matrices to obtain better compression rate by keeping lower rank for unimportant neurons.…”
Section: Related Workmentioning
confidence: 99%
“…Low-rank decomposition: Low-rank decomposition algorithms [ 30 , 31 , 32 ] use a lower-rank set instead of the original set of parameters to approximate the CNN to achieve compression. Swaminathan et al [ 31 ] argue that the low-rank decomposition of weight matrices should consider influence of both input as well as output neurons of a layer. They propose a sparse low rank (SLR) approach that sparsifies SVD matrices to obtain better compression rate by keeping lower rank for unimportant neurons.…”
Section: Related Workmentioning
confidence: 99%
“…The parameter pruning, hashing and quantisation methods explore model parameter redundancy and seek to remove redundant and uncritical parameters. Low-ranking factorisation based techniques [40], [41] employ matrix/tensor decomposition to estimate informative parameters of the DNNs. Convolutional filters based on methods utilise compact/transferred convolutional techniques design specially-tailored convolutional filters to reduce the parameter space and, to save storage and computation.…”
Section: Active Research Problems In Deep Learningmentioning
confidence: 99%
“…Effective methods for reducing the size of models and the computation parameters include the use of information compression methods such as SqueezeNet [27] and depth-wise separable filters like the MobileNets [28]. Post-training model optimization can instead be achieved without important loss of performance, by employing techniques like quantization [29], factorization [30], distillation [31] and pruning [32]. Edge efficient models development has recently led to an industry movement toward such a framework.…”
Section: Introductionmentioning
confidence: 99%