2017 International Joint Conference on Neural Networks (IJCNN) 2017
DOI: 10.1109/ijcnn.2017.7966420
|View full text |Cite
|
Sign up to set email alerts
|

Compressing recurrent neural network with tensor train

Abstract: Recurrent Neural Network (RNN) are a popular choice for modeling temporal and sequential tasks and achieve many state-of-the-art performance on various complex problems. However, most of the state-of-the-art RNNs have millions of parameters and require many computational resources for training and predicting new data. This paper proposes an alternative RNN model to reduce the number of parameters significantly by representing the weight parameters based on Tensor Train (TT) format. In this paper, we implement … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
62
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 86 publications
(62 citation statements)
references
References 23 publications
0
62
0
Order By: Relevance
“…Different matrix decomposition techniques can be also related to this class of compression methods. These methods can be as simple as low-rank decomposition or more complex like Tensor Train (TT) decomposition [24,25,26,27]. However, the TT-based approach have not been studied in language modeling task, where there are such issues as highdimensional input and output data, and, as a consequence, more options to configure TT decomposition.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Different matrix decomposition techniques can be also related to this class of compression methods. These methods can be as simple as low-rank decomposition or more complex like Tensor Train (TT) decomposition [24,25,26,27]. However, the TT-based approach have not been studied in language modeling task, where there are such issues as highdimensional input and output data, and, as a consequence, more options to configure TT decomposition.…”
Section: Related Workmentioning
confidence: 99%
“…This approach was successfully applied to compress fully connected neural networks [24], to develop convolutional TT layer [25] and to compress and improve RNNs [26,27]. However, there are still no studies of the TT decomposition for language modeling and similar tasks with high-dimensional outputs at the softmax layer.…”
Section: Tensor Train Decompositionmentioning
confidence: 99%
“…A common challenge of the above technique is to determine the tensor rank. Exactly determining a tensor rank in general [49]- [51] Tensor Train [49] CP [55] Tucker [53], [55], [57] Tensor Train [52], [54], [55] CP [48] Tucker Tensor Train is NP-hard [47]. Therefore, in practice one often leverages numerical optimization or statistical techniques to obtain a reasonable rank estimation.…”
Section: Compact Deep Learning Modelsmentioning
confidence: 99%
“…In order to avoid the expensive pre-training in the uncompressed format, the work in [52] and [53] directly trained fully connected and convolution layers in low-rank tensor-train and Tucker format with the tensor ranks fixed in advance. This idea has also been applied to recurrent neural networks [54], [55].…”
Section: Tensorized Training With a Fixed Rankmentioning
confidence: 99%
“…There are a lot of RNNs compression methods based on specific weight matrix representations (Tjandra et al, 2017;Le et al, 2015) or sparsification (Narang et al, 2017;Wen et al, 2018). In this paper we focus on RNNs compression via sparsification.…”
Section: Introductionmentioning
confidence: 99%