2002
DOI: 10.1006/jcss.2001.1795
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive and Self-Confident On-Line Learning Algorithms

Abstract: We study on-line learning in the linear regression framework. Most of the performance bounds for on-line algorithms in this framework assume a constant learning rate. To achieve these bounds the learning rate must be optimized based on a posteriori information. This information depends on the whole sequence of examples and thus it is not available to any strictly on-line algorithm. We introduce new techniques for adaptively tuning the learning rate as the data sequence is progressively revealed. Our techniques… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

9
207
0

Year Published

2003
2003
2024
2024

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 154 publications
(216 citation statements)
references
References 35 publications
9
207
0
Order By: Relevance
“…A simple algebraic manipulation of the above implies the following theorem We can optimize for β in advance, or do it dynamically using Auer et al (2002b) …”
Section: Where L Hi = ∑ T I(t) T H Is the Loss Of The Online Algoritmentioning
confidence: 99%
“…A simple algebraic manipulation of the above implies the following theorem We can optimize for β in advance, or do it dynamically using Auer et al (2002b) …”
Section: Where L Hi = ∑ T I(t) T H Is the Loss Of The Online Algoritmentioning
confidence: 99%
“…The algorithm used to update the weights of the second layer is the Incrementally Adaptive Weighted Majority (IAWM) [15]. Denote with o …”
Section: B Implementation and Analysismentioning
confidence: 99%
“…We refer the reader to [15] for the exact details of the update rule of IAWM. Chaining the bound of the PA with the bound of IAWM we have that…”
Section: B Implementation and Analysismentioning
confidence: 99%
“…The tuning parameter η can be set optimally only when the time length n is known in advance. However, we recall a simple modification of the exponentially weighted average algorithm, proposed by Auer, Cesa-Bianchi, and Gentile (2002), which does not need to know n in advance. A natural adaptive version of the optimal parameter η determined in the case of known time length is formed by defining the tuning parameter at round t by η t = B −1 8 ln N/t.…”
Section: Sequential Prediction: External Regretmentioning
confidence: 99%