Advances in Neural Information Processing Systems 14 2002
DOI: 10.7551/mitpress/1120.003.0051
|View full text |Cite
|
Sign up to set email alerts
|

On the Generalization Ability of On-line Learning Algorithms

Abstract: In this paper, it is shown how to extract a hypothesis with small risk from the ensemble of hypotheses generated by an arbitrary on-line learning algorithm run on an independent and identically distributed (i.i.d.) sample of data. Using a simple large deviation argument, we prove tight data-dependent bounds for the risk of this hypothesis in terms of an easily computable statistic associated with the on-line performance of the ensemble. Via sharp pointwise bounds on , we then obtain risk tail bounds for kernel… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
12
0

Year Published

2006
2006
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 25 publications
(13 citation statements)
references
References 44 publications
1
12
0
Order By: Relevance
“…The maximum error was chosen as the performance metric to emphasize reducing the error of outlying predictions, by grading the predictive ability of the model on its worst prediction. The choice of Bayesian optimization was based on its favorable comparison with similar methods, such as the grid search method (24).…”
Section: Optimization Of Hyperparameterssupporting
confidence: 93%
“…The maximum error was chosen as the performance metric to emphasize reducing the error of outlying predictions, by grading the predictive ability of the model on its worst prediction. The choice of Bayesian optimization was based on its favorable comparison with similar methods, such as the grid search method (24).…”
Section: Optimization Of Hyperparameterssupporting
confidence: 93%
“…Previous latent state models of behaviour have concentrated on animal behaviour in relatively well-studied task environments, such as courtship behaviours in fruit flies [17], visual detection in mice [16, 26, 18], or swim bouts in larval zebrafish [62]. These paradigms lend themselves well to models of latent states as the resultant observations are intuitively discretisable.…”
Section: Discussionmentioning
confidence: 99%
“…The probabilistic choice is sometimes known as a "softmax" approach in reinforcement learning studies (Cesa-Bianchi et al, 2017) and the parameter α is a strength-of-response parameter. If α= 0 response is random, and as α grows the probability of choosing the action a with the highest r t a ð Þ increases.…”
Section: The Neural Autopilot Modelmentioning
confidence: 99%