2018
DOI: 10.48550/arxiv.1807.02291
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Sliced Recurrent Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 2 publications
0
7
0
Order By: Relevance
“…For recurrent neural networks, the more advanced architectures help people deal with the gradient vanishing or explosion issue and accelerate the execution of RNN. However, how to parallelize RNN is still a big problem under active investigation [187].…”
Section: State-of-the-art Deep Architecturesmentioning
confidence: 99%
“…For recurrent neural networks, the more advanced architectures help people deal with the gradient vanishing or explosion issue and accelerate the execution of RNN. However, how to parallelize RNN is still a big problem under active investigation [187].…”
Section: State-of-the-art Deep Architecturesmentioning
confidence: 99%
“…Reference [24] presented an effective approach to train RNN on multiple GPUs, where parallelized stochastic gradient descent (SGD) was applied and achieved 3.4 times speedup on 4 GPUs than the single one. More recently, [25] introduced sliced recurrent neural networks (SRNNs), which could be parallelized by slicing the sequences into many subsequences. SRNNs have the ability to obtain high-level information through multiple layers with few extra parameters.…”
Section: Related Workmentioning
confidence: 99%
“…Fast Inference AQM+'s time complexity can be decreased further by changing the structure of aprxAgen. In specific, we can apply diverse methods such as skipping the update of hidden states in some steps , using convolution networks or self-attention networks Vaswani et al, 2017), substituting matrix multiplication operation for hidden state update to weighted addition (Yu & Liu, 2018), and direct information gain inference from the neural networks (Belghazi et al, 2018).…”
Section: Toward Practical Applicationsmentioning
confidence: 99%