2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00977
|View full text |Cite
|
Sign up to set email alerts
|

Learning Compact Recurrent Neural Networks with Block-Term Tensor Decomposition

Abstract: Recurrent Neural Networks (RNNs) are powerful sequence modeling tools. However, when dealing with high dimensional inputs, the training of RNNs becomes computational expensive due to the large number of model parameters. This hinders RNNs from solving many important computer vision tasks, such as Action Recognition in Videos and Image Captioning. To overcome this problem, we propose a compact and flexible structure, namely Block-Term tensor decomposition, which greatly reduces the parameters of RNNs and improv… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
79
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
3
3
3

Relationship

1
8

Authors

Journals

citations
Cited by 110 publications
(79 citation statements)
references
References 43 publications
0
79
0
Order By: Relevance
“…Based on the empirical results, TT-format are able to reduce the number of parameters significantly and retain the model performance at the same time. Recent work from [36] used block decompositions to represent the RNN weight matrices.…”
Section: Resultsmentioning
confidence: 99%
“…Based on the empirical results, TT-format are able to reduce the number of parameters significantly and retain the model performance at the same time. Recent work from [36] used block decompositions to represent the RNN weight matrices.…”
Section: Resultsmentioning
confidence: 99%
“…Our study concentrated on the ability to estimate dose in heterogeneous geometries, and no effort was made in improving the model efficiency. Various model compression techniques, for example, pruning, quantization, and tensor decomposition methods (achieving low-rank structures in the weight matrices), [51][52][53] may substantially lower the number of parameters in fully connected layers. 54,55 The efficiency of the model can be further enhanced through fine-tuning of the model architecture.…”
Section: In This Paper We Have Demonstrated the General Feasibility mentioning
confidence: 99%
“…-Low rank factorization: [10,36] -Factorized embedding parameterization: [19] -Block-Term tensor decomposition: [23,38] -Singular Value Decomposition: [37] -Joint factorization of recurrent and inter-layer weight matrices: [28] -Tensor train decomposition: [10,17] -Sparse factorization: [6] • [11] • Applications: In this section, we will discuss application and success of various model compression methods across various popular NLP tasks like Language modeling, Machine translation, Summarization, Sentiment analysis, Question answering, Natural language inference, Paraphrasing, Image captioning, Handwritten character recognition. • Summary and future trends.…”
Section: Tutorial Outlinementioning
confidence: 99%