Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.436
|View full text |Cite
|
Sign up to set email alerts
|

Tensorized Embedding Layers

Abstract: The embedding layers transforming input words into real vectors are the key components of deep neural networks used in natural language processing. However, when the vocabulary is large, the corresponding weight matrices can be enormous, which precludes their deployment in a limited resource setting. We introduce a novel way of parameterizing embedding layers based on the Tensor Train decomposition, which allows compressing the model significantly at the cost of a negligible drop or even a slight gain in perfo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 20 publications
(9 citation statements)
references
References 24 publications
0
8
0
Order By: Relevance
“…Matrix decomposition (e.g., ALBERT (Lan et al, 2019) in the embedding layer and (Noach & Goldberg, 2020)) could decrease parameter scale with a linear factor depending on the selected rank. More advanced tensor decomposition approaches can be implemented by tensor network, which has recently been used to compress general neural networks (Gao et al, 2020;Novikov et al, 2015), compress embedding layer (Khrulkov et al, 2019;Hrinchuk et al, 2020;Panahi et al, 2019).…”
Section: Methodsmentioning
confidence: 99%
“…Matrix decomposition (e.g., ALBERT (Lan et al, 2019) in the embedding layer and (Noach & Goldberg, 2020)) could decrease parameter scale with a linear factor depending on the selected rank. More advanced tensor decomposition approaches can be implemented by tensor network, which has recently been used to compress general neural networks (Gao et al, 2020;Novikov et al, 2015), compress embedding layer (Khrulkov et al, 2019;Hrinchuk et al, 2020;Panahi et al, 2019).…”
Section: Methodsmentioning
confidence: 99%
“…Among the promising directions, we should mention computations with reduced precision, approximate methods [Novikov et al, 2022], randomized computations , and structured NN layers [Hrinchuk et al, 2020], including those based on tensor factorizations. We should highlight the importance for these approximate methods to be additive in the sense that they can be combined and still provide sufficient enough performance with reasonable quality degradation.…”
Section: Conclusion and Further Researchmentioning
confidence: 99%
“…(Chen et al, 2018) proposed the blockwise low-rank approximation method for word embedding. (Hrinchuk et al, 2020) devised a way of interpreting an embedding matrix into a 3-dimensional tensor and proposed an embedding structure by decomposing it with tensor-train decomposition. (Panahi et al, 2020) proposed a smallsize word embedding structure inspired by quantum entanglement.…”
Section: Related Workmentioning
confidence: 99%
“…TensorTrain is the tensor-train decomposedbased method in (Hrinchuk et al, 2020). In (Hrinchuk et al, 2020), TensorTrain is computed by training from scratch.…”
Section: Implementation Detailsmentioning
confidence: 99%
See 1 more Smart Citation