2005
DOI: 10.1007/11503415_15
|View full text |Cite
|
Sign up to set email alerts
|

Improved Second-Order Bounds for Prediction with Expert Advice

Abstract: This work studies external regret in sequential prediction games with both positive and negative payoffs. External regret measures the difference between the payoff obtained by the forecasting strategy and the payoff of the best action. In this setting, we derive new and sharper regret bounds for the well-known exponentially weighted average forecaster and for a new forecaster with a different multiplicative update rule. Our analysis has two main advantages: first, no preliminary knowledge about the payoff seq… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
206
0

Year Published

2005
2005
2014
2014

Publication Types

Select...
5
1

Relationship

2
4

Authors

Journals

citations
Cited by 100 publications
(207 citation statements)
references
References 15 publications
1
206
0
Order By: Relevance
“…For example, using the algorithms of Cesa-Bianchi et al (2005) one can get a more refined regret bound, which depends on the second moment.…”
Section: Corollary 6 Using An Optimized Experts Algorithm As the A I mentioning
confidence: 99%
See 1 more Smart Citation
“…For example, using the algorithms of Cesa-Bianchi et al (2005) one can get a more refined regret bound, which depends on the second moment.…”
Section: Corollary 6 Using An Optimized Experts Algorithm As the A I mentioning
confidence: 99%
“…We need to use here an external regret algorithm which does not need to have as an input the value of L i min . An example of such an algorithm is Corollary 2 in Cesa- Bianchi et al (2005), which guarantees an external regret of at most O( √ L min log N + log N).…”
Section: Lower Bounds On Swap Regretmentioning
confidence: 99%
“…These quantities are of importance, because they are used in schemes for adaptively tuning the learning rate online. In particular, [6] introduces a parameter-free online tuning scheme based on the variance, for which the expected regret is at most of the order…”
Section: Computing Expected Loss and Variancementioning
confidence: 99%
“…However, the overall running time changes from O (T ) to O (T 2 ). 6 We don't know how to do the fancier method efficiently, which mixes in a bit of the past average distribution. The reason is that the exponential weight updates on the combined lists seem to be at loggerheads with mixing in the past average weight.…”
Section: Open Problemsmentioning
confidence: 99%
“…al. [7] studied second-order bounds for exponentially weighted average forecaster and they analyzed the expected regret of the algorithm in the full monitoring case when the bound of the loss function unknown. They indicated their results in partial monitoring case.…”
Section: Introductionmentioning
confidence: 99%