2022
DOI: 10.48550/arxiv.2203.02617
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

How to Train Unstable Looped Tensor Network

Abstract: A rising problem in the compression of Deep Neural Networks is how to reduce the number of parameters in convolutional kernels and the complexity of these layers by low-rank tensor approximation. Canonical polyadic tensor decomposition (CPD) and Tucker tensor decomposition (TKD) are two solutions to this problem and provide promising results. However, CPD often fails due to degeneracy, making the networks unstable and hard to fine-tune. TKD does not provide much compression if the core tensor is big. This moti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 38 publications
(57 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?