2007
DOI: 10.1162/neco.2007.19.3.757
|View full text |Cite
|
Sign up to set email alerts
|

Training Recurrent Networks by Evolino

Abstract: In recent years, gradient-based LSTM recurrent neural networks (RNNs) solved many previously RNN-unlearnable tasks. Sometimes, however, gradient information is of little use for training RNNs, due to numerous local minima. For such cases, we present a novel method: EVOlution of systems with LINear Outputs (Evolino). Evolino evolves weights to the nonlinear, hidden nodes of RNNs while computing optimal linear mappings from hidden state to output, using methods such as pseudoinverse-based linear regression. If w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
107
0
2

Year Published

2009
2009
2016
2016

Publication Types

Select...
7
2

Relationship

1
8

Authors

Journals

citations
Cited by 201 publications
(115 citation statements)
references
References 28 publications
1
107
0
2
Order By: Relevance
“…A link has often been made between Reservoir Computing and kernel machines (Schmidhuber et al, 2007;Shi and Han, 2007), since both techniques essentially map the input data to a high-dimensional space, called feature-space, in which classification or regression is then performed linearly. In the case of reservoirs, this mapping is performed explicitly, as the hidden state of the reservoir is mapped directly onto the output.…”
Section: Introductionmentioning
confidence: 99%
“…A link has often been made between Reservoir Computing and kernel machines (Schmidhuber et al, 2007;Shi and Han, 2007), since both techniques essentially map the input data to a high-dimensional space, called feature-space, in which classification or regression is then performed linearly. In the case of reservoirs, this mapping is performed explicitly, as the hidden state of the reservoir is mapped directly onto the output.…”
Section: Introductionmentioning
confidence: 99%
“…• y 2 : MSE=3.57E-8, NRMSE=1.88E-4, using 13 SV; • y 3 : MSE=2.33E-6, NRMSE=1.23E-3, using 15 SV; • y 4 : MSE=1.75E-5, NRMSE=2.86E-3, using 15 SV; • y 5 : MSE=9.16E-5, NRMSE=5.94E-3, using 15 SV, which are considerably better than the errors reported in (Xue, Yang & Haykin, 2007) and (Schmidhuber, Wierstra, Gagliolo & Gomez, 2007) but still the orders of magnitude worse than the errors obtained by the corresponding optimized SOPNN-LM model shown in Table 3. Also the computational complexity (Maric & Ivek, 2011) of the SVM model with 13 or 15 support vectors is much higher than the complexity of the corresponding SOPNN-LM model with 3, 8, 13 or 16 2-dimensional, 2 nd order node-polynomials (1).…”
Section: Tablementioning
confidence: 86%
“…We analyze the performances of SOPNN on multiple superimposed oscillations (MSO) task that has already been attempted with varying degrees of success by using ANN and SVM (Xue, Yang & Haykin, 2007;Schmidhuber, Wierstra, Gagliolo & Gomez, 2007;Holzmann & Hauser, 2010;Ceperic, Gielen & Baric, 2012). The following multiple sinusoids are modeled: where n=1,…,700.…”
Section: Multiple Superimposed Oscillations Modelingmentioning
confidence: 99%
“…There exist reservoir computing networks that make use of evolutionary computation for training the weights of the reservoir, such as Evolino [144], and several other models are currently proposed, with or without spiking neurons [40,76,75]. Although the research area is in rapid expansion, several papers [175,151,94] propose valuable surveys.…”
Section: Related Reservoir Computing Workmentioning
confidence: 99%