1997
DOI: 10.1088/0954-898x_8_3_004
|View full text |Cite
|
Sign up to set email alerts
|

Optimal ensemble averaging of neural networks

Abstract: Based on an observation about the different effect of ensemble averaging on the bias and variance portions of the prediction error, we discuss training methodologies for ensembles of networks. We demonstrate the effect of variance reduction and present a method of extrapolation to the limit of an infinite ensemble. A significant reduction of variance is obtained by averaging just over initial conditions of the neural networks, without varying architectures or training sets. The minimum of the ensemble predicti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

1
92
0
3

Year Published

2001
2001
2023
2023

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 157 publications
(97 citation statements)
references
References 12 publications
1
92
0
3
Order By: Relevance
“…The second sum contains the cross terms of the ensemble members and disappears if the models are completely uncorrelated [8]. So the reduction in the variance of the ensemble is related to the degree of independence of the single models [7]. This is the key feature of the ensemble approach.…”
Section: Ensemblesmentioning
confidence: 99%
See 1 more Smart Citation
“…The second sum contains the cross terms of the ensemble members and disappears if the models are completely uncorrelated [8]. So the reduction in the variance of the ensemble is related to the degree of independence of the single models [7]. This is the key feature of the ensemble approach.…”
Section: Ensemblesmentioning
confidence: 99%
“…The extension to classification problems was straightforward after the formulation of a bias-variance decomposition for zero-one loss functions [4], [5]. The key feature of the ensemble approach is the introduction of model diversity [6], [7], [8] that helps to reduce the variance of the resulting ensemble model. There are several ways to achieve diverse models like the well known bootstrap aggregating or 'bagging' (see Breiman [9]) where the models are trained on different subsets of the training data or heterogeneous ensembles, that consist of several different model classes like Neural Networks, nearest-neighbor models, decision trees, etc [10], [11].…”
Section: Introductionmentioning
confidence: 99%
“…Some of the ensemble models are built and implemented as follows. In 1997, Naftaly, Intrator, and Horn revealed that the ensemble averaging was a powerful procedure which, when used correctly, improved on single network performance [12]. In 2004, the ensemble of various neural networks consisted of MLPN, Elman recurrent neural network, radial basis function network and Hopfield model was proposed for building a r obust weather forecasting [13].…”
Section: Introductionmentioning
confidence: 99%
“…The approach is not very reliable because the result is so sensitive to the estimated model parameters. The main purpose of the present letter is to introduce the idea of ensemble technique [Hansen & Salamon, 1990;Meir, 1995;Krogh & Sollich, 1997;Naftaly et al, 1997] in order to realize a robust reconstruction of the synchronization diagrams from time series data.…”
Section: Introductionmentioning
confidence: 99%