2004
DOI: 10.1162/08997660460734010
|View full text |Cite
|
Sign up to set email alerts
|

On the Asymptotic Distribution of the Least-Squares Estimators in Unidentifiable Models

Abstract: In order to analyze the stochastic property of multilayered perceptrons or other learning machines, we deal with simpler models and derive the asymptotic distribution of the least-squares estimators of their parameters. In the case where a model is unidentified, we show different results from traditional linear models: the well-known property of asymptotic normality never holds for the estimates of redundant parameters.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2007
2007
2012
2012

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 18 publications
0
4
0
Order By: Relevance
“…In regular statistical models, Bayes a posteriori distribution converges to the normal distribution and the maximum likelihood estimator satisfies asymptotic normality. Whereas, in singular learning machines, Bayes a posteriori distribution converges to the singular distribution [17] and the maximum likelihood estimator diverges to infinity [6], [5], [7]. Singularities in the parameter space strongly affect learning dynamics [1].…”
Section: A Backgroundmentioning
confidence: 99%
“…In regular statistical models, Bayes a posteriori distribution converges to the normal distribution and the maximum likelihood estimator satisfies asymptotic normality. Whereas, in singular learning machines, Bayes a posteriori distribution converges to the singular distribution [17] and the maximum likelihood estimator diverges to infinity [6], [5], [7]. Singularities in the parameter space strongly affect learning dynamics [1].…”
Section: A Backgroundmentioning
confidence: 99%
“…In such learning machines, the map taking parameters to probability distributions is not one-to-one and the Fisher information matrices are singular, hence they are called singular learning machines. For example, three-layered neural networks, normal mixtures, hidden Markov models, Bayesian networks, and reduced rank regressions are singular learning machines [1,2,4,5,6,10]. If a statistical model is singular, then either the maximum likelihood estimator is not subject to the normal distribution even asymptotically or the Bayes posterior distribution can not be approximated by any normal distribution.…”
Section: Introductionmentioning
confidence: 99%
“…For singular learning machines, the log likelihood function can not be approximated by any quadratic form of the parameter, with the result that the conventional relationship between generalization errors and training errors does not hold either for the maximum likelihood method [6] [5] [7] or Bayes estimation [12]. Singularities strongly affect generalization performances [15] and learning dynamics [1].…”
Section: Introductionmentioning
confidence: 99%