2009
DOI: 10.1142/s0218001409007132
|View full text |Cite
|
Sign up to set email alerts
|

An Incremental Framework Based on Cross-Validation for Estimating the Architecture of a Multilayer Perceptron

Abstract: We define the problem of optimizing the architecture of a multilayer perceptron (MLP) as a state space search and propose the MOST (Multiple Operators using Statistical Tests) framework that incrementally modifies the structure and checks for improvement using cross-validation. We consider five variants that implement forward/backward search, using single/multiple operators and searching depth-first/breadth-first. On 44 classification and 30 regression datasets, we exhaustively search for the optimal and evalu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0
1

Year Published

2010
2010
2016
2016

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 18 publications
(8 citation statements)
references
References 57 publications
0
7
0
1
Order By: Relevance
“…Two model selection approaches based on a K-fold crossvalidation (K-CV) model evaluation methodology, which is commonly used in the literature [14,[24][25][26], are also applied for comparison purposes. The main differences with respect to MoSe, MoSe-R and MoSe-D model selection methodologies are in terms of the generation of the training and validation sets, the RBFNN optimization process and the iterative procedure typical of a cross-validation approach to evaluate a given model/structure network (Fig.…”
Section: Strategies Based On a K-fold Cross-validation Methodologymentioning
confidence: 99%
See 3 more Smart Citations
“…Two model selection approaches based on a K-fold crossvalidation (K-CV) model evaluation methodology, which is commonly used in the literature [14,[24][25][26], are also applied for comparison purposes. The main differences with respect to MoSe, MoSe-R and MoSe-D model selection methodologies are in terms of the generation of the training and validation sets, the RBFNN optimization process and the iterative procedure typical of a cross-validation approach to evaluate a given model/structure network (Fig.…”
Section: Strategies Based On a K-fold Cross-validation Methodologymentioning
confidence: 99%
“…A network smaller than the optimal architecture underfits and fails to learn the data well (bias is high and variance is low), and a large network suffers the overfitting problem, resulting in poor generalization (bias is low and variance is high). Thus, the optimal architecture is the one with low bias and low variance so that the network learns the function underlying the data and not the existing noise [14].…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…The other common approaches for optimizing neural network architecture are basically growing, pruning and a combination of two strategies namely growing and pruning [13]. The first, also called as constructive methods, start with a minimal network and add new hidden units during the training process [14][15][16].…”
Section: Neural Network Architecturementioning
confidence: 99%