2019
DOI: 10.48550/arxiv.1909.09586
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Understanding LSTM -- a tutorial into Long Short-Term Memory Recurrent Neural Networks

Abstract: Long Short-Term Memory Recurrent Neural Networks (LSTM-RNN) are one of the most powerful dynamic classifiers publicly known. The network itself and the related learning algorithms are reasonably well documented to get an idea how it works. This paper will shed more light into understanding how LSTM-RNNs evolved and why they work impressively well, focusing on the early, ground-breaking publications. We significantly improved documentation and fixed a number of errors and inconsistencies that accumulated in pre… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
100
0
2

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 171 publications
(188 citation statements)
references
References 36 publications
0
100
0
2
Order By: Relevance
“…They are composed of LSTM cells capable of capturing long-term dependencies in sequences while attenuate gradient vanishing/exploding problem [28]. This capacity is achieved by the use of forget and update gates to modify memory cell state that allow gradients to also flow unchanged [29,30]. The LSTM memory cells are composed by self-loops that encoded temporal information in the cell states, and three regulators gates that operate the flow of information within each cell.…”
Section: Long Short Term Memorymentioning
confidence: 99%
See 1 more Smart Citation
“…They are composed of LSTM cells capable of capturing long-term dependencies in sequences while attenuate gradient vanishing/exploding problem [28]. This capacity is achieved by the use of forget and update gates to modify memory cell state that allow gradients to also flow unchanged [29,30]. The LSTM memory cells are composed by self-loops that encoded temporal information in the cell states, and three regulators gates that operate the flow of information within each cell.…”
Section: Long Short Term Memorymentioning
confidence: 99%
“…The three gates are called: forget gate f g , input gate i g and output gate o g , which operate the information flow by erasing, writing and reading, respectively. Therefore, LSTM models memorize information at different intervals and are suitable to predict time series with a certain duration interval [30,31].…”
Section: Long Short Term Memorymentioning
confidence: 99%
“…jects. [14] regards the common occurrence frequency of object pairs as prior knowledge and utilizes LSTM (Long shortterm memory) [15] as an encoder to transfer context information to improve the feature representation between objects.…”
Section: -mentioning
confidence: 99%
“…To train the CG agent we subdivide its architecture into a block which estimates the gradient, and into another block that controls an internal memory M (t i ) of the chemical field (i.e., the chemical memory control cell (CMC)). The latter is inspired by the well-known long short-term memory (LSTM) cell [39,40]. The first block is trained using the NEAT algorithm: it takes as input L T (t i ) and V T (t i ) as well as two recurrent variables C x (t i ) and G x (t i ) and maps this information onto a control output C y (t i ) and an estimated value of the instantaneous chemical gradient G y (t i ) ∈ [−1, 1], both to be forwarded to the CMC cell.…”
Section: Phase One: Learning Unidirectional Locomotionmentioning
confidence: 99%