2023
DOI: 10.1007/978-1-0716-3195-9_4
|View full text |Cite
|
Sign up to set email alerts
|

Recurrent Neural Networks (RNNs): Architectures, Training Tricks, and Introduction to Influential Research

Susmita Das,
Amara Tariq,
Thiago Santos
et al.

Abstract: Recurrent neural networks (RNNs) are neural network architectures with hidden state and which use feedback loops to process a sequence of data that ultimately informs the final output. Therefore, RNN models can recognize sequential characteristics in the data and help to predict the next likely data point in the data sequence. Leveraging the power of sequential data processing, RNN use cases tend to be connected to either language models or time-series data analysis. However, multiple popular RNN architectures… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(4 citation statements)
references
References 21 publications
0
4
0
Order By: Relevance
“…RNNs characterize a specialized class of ANNs engineered to manage sequential data [ 49 , 50 ]. They achieve this by incorporating connections that create directed cycles within the network graph, facilitating dynamic temporal behavior and the processing of sequences of variable lengths.…”
Section: Types Of Pnnsmentioning
confidence: 99%
“…RNNs characterize a specialized class of ANNs engineered to manage sequential data [ 49 , 50 ]. They achieve this by incorporating connections that create directed cycles within the network graph, facilitating dynamic temporal behavior and the processing of sequences of variable lengths.…”
Section: Types Of Pnnsmentioning
confidence: 99%
“…Traditional RNNs suffer from the vanishing gradient problem, which makes it difficult for the model to capture long-range dependencies in a sequence. LSTM is an advancement over traditional RNNs, designed to overcome the vanishing gradient problem by introducing memory cells, input gates, forget gates and output gates to control the flow of information into and out of the cells [91]. The RNN model can transfer information more efficiently by using skip connections.…”
Section: Recurrent Neural Network (Rnn)mentioning
confidence: 99%
“…The reset gate regulates the extent to which past information must be forgotten to update the relevant memory, while the update gate regulates the retention or abandonment of information and the integration of new information from input into existing memory. With these two gates, GRU provides adaptive control over the flow of information in recurrent networks [91]. Marzinotto et al [56] use four layers of bidirectional GRU and are equipped with various more complex features for SRL tasks, including morphology, surface characteristics, and syntax.…”
Section: Recurrent Neural Network (Rnn)mentioning
confidence: 99%
“…These crucial state transitions are meticulously orchestrated by a set of adaptive gating units, which include the input, forget, and output gates, each performing a specific regulatory function to ensure the fidelity of information flow across the temporal expanse of the sequence. There are three kinds of gates in the LSTM layer, input gate, forget gate, and output gate (35). Figure 1 illustrates the flow of data at time step t and shows how the gates forget, update, and output the cell and hidden states.…”
Section: Constructing the Lstm Model 241 Cell Structure Of Lstm Networkmentioning
confidence: 99%