Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks. IJCNN 2000. Neural Computing: New Challeng 2000
DOI: 10.1109/ijcnn.2000.859438
|View full text |Cite
|
Sign up to set email alerts
|

On the problem in model selection of neural network regression in overrealizable scenario

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0
1

Year Published

2001
2001
2012
2012

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(11 citation statements)
references
References 22 publications
0
10
0
1
Order By: Relevance
“…We may also consider using likelihood-based model selection criteria, AIC or BIC, which have been widely used for various types of selection problems. However, derivations of these criteria are invalid for mixture models because of the identification problem (Hagiwara et al, 2000;Miloslavsky and Van der Laan, 2003). Nevertheless, because of the computational simplicity, they are still considered as popular model selection criteria for the selection of the number of groups (Miloslavsky and Van der Laan, 2003).…”
Section: Model Comparisons All Competing Models Inmentioning
confidence: 99%
“…We may also consider using likelihood-based model selection criteria, AIC or BIC, which have been widely used for various types of selection problems. However, derivations of these criteria are invalid for mixture models because of the identification problem (Hagiwara et al, 2000;Miloslavsky and Van der Laan, 2003). Nevertheless, because of the computational simplicity, they are still considered as popular model selection criteria for the selection of the number of groups (Miloslavsky and Van der Laan, 2003).…”
Section: Model Comparisons All Competing Models Inmentioning
confidence: 99%
“…Hartigan has suggested by a sophisticated technique that the likelihood ratio is not of the order of 1/n but of the order of log log n/n in the case of the Gaussian mixture model [13]. Hagiwara's discussion is also a product of elaborate work [12]. Dacunha-Castelle and Gassiat [8] have developed the general framework of this problem.…”
Section: What Happens In Singular Modelsmentioning
confidence: 99%
“…Hagiwara and colleagues have performed simulations and shown that the AIC does not work on neural network models called multilayer perceptrons; they insist that this is caused by a hierarchical property of the model. Using simple models, they have shown that the least square error of the estimator does not obey asymptotically the usual rule of 1/n (n is the number of data) but instead the law of log n/n [12]. Kitahara and colleagues have obtained similar results for different models [15].…”
Section: Introductionmentioning
confidence: 99%
“…In such learning machines, the map taking parameters to probability distributions is not one-to-one and the Fisher information matrices are singular, hence they are called singular learning machines. For example, three-layered neural networks, normal mixtures, hidden Markov models, Bayesian networks, and reduced rank regressions are singular learning machines [1,2,4,5,6,10]. If a statistical model is singular, then either the maximum likelihood estimator is not subject to the normal distribution even asymptotically or the Bayes posterior distribution can not be approximated by any normal distribution.…”
Section: Introductionmentioning
confidence: 99%