2018
DOI: 10.3390/info9030056
|View full text |Cite
|
Sign up to set email alerts
|

Residual Recurrent Neural Networks for Learning Sequential Representations

Abstract: Recurrent neural networks (RNN) are efficient in modeling sequences for generation and classification, but their training is obstructed by the vanishing and exploding gradient issues. In this paper, we reformulate the RNN unit to learn the residual functions with reference to the hidden state instead of conventional gated mechanisms such as long short-term memory (LSTM) and the gated recurrent unit (GRU). The residual structure has two main highlights: firstly, it solves the gradient vanishing and exploding is… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
45
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 69 publications
(48 citation statements)
references
References 25 publications
0
45
0
Order By: Relevance
“…In the case when this process takes place, due to the fact that many derivatives having low-values are multiplied when computing the chain rule, the gradient vanishes to 0. In the case of the LSTM ANNs, the derivative of the identical function from Equation 2is the constant function 1 and this fact represents a certain advantage in the case when the LSTM is trained based on the backpropagation, because in this case the gradient is not vanishing [47].…”
Section: The Long Short-term Memory (Lstm) Neural Networkmentioning
confidence: 99%
“…In the case when this process takes place, due to the fact that many derivatives having low-values are multiplied when computing the chain rule, the gradient vanishes to 0. In the case of the LSTM ANNs, the derivative of the identical function from Equation 2is the constant function 1 and this fact represents a certain advantage in the case when the LSTM is trained based on the backpropagation, because in this case the gradient is not vanishing [47].…”
Section: The Long Short-term Memory (Lstm) Neural Networkmentioning
confidence: 99%
“…Numerical data modeling is a broad research area which involves a wide variety of novel artificial-intelligence and statistical tools. A few examples of recent contributions in this field are Petri nets to model honeypot [13], extreme learning machines in monthly precipitation time series forecasting [14], and recurrent neural networks to language modeling, emotion classification and polyphonic modeling [15]. Specifically to the problem treated in this paper, traditional artificial neural-network techniques to build up empirical models of GFR were proposed and tested in the scientific literature, although their performance appeared unsatisfactory or questionable [16,17].…”
Section: Introductionmentioning
confidence: 99%
“…To conclude this section of the paper, let us consider a minimal working numerical example. Let D X = [2, 3, 5, 5, 6, 6, 7, 9] and D Y =[6,7,10,10,11,11,11,12,15,20], i.e., n = 8, m = 10. The actual underlying model is f (x) = 2x + 2.…”
mentioning
confidence: 99%
“…In this section, we evaluate the proposed prediction model, denoted as URL-LSTM, with other baseline algorithms, including Random Forest Regression (RF) [17], Support Vector Regression SVR [18], Long Short Term Memory (LSTM) [19] and Residual Recurrent Networks (RRN) [20]. We do not compare our proposed method with other power consumption prediction approaches in which intrusive features are needed.…”
Section: Experimental Evaluationsmentioning
confidence: 99%