2015
DOI: 10.1007/s10994-015-5531-y
|View full text |Cite
|
Sign up to set email alerts
|

Lower bounds on individual sequence regret

Abstract: In this work we lower bound the individual sequence anytime regret of a large family of online algorithms. This bound depends on the quadratic variation of the sequence, Q T , and the learning rate. Nevertheless, we show that any learning rate that guarantees a regret upper bound of O( √ Q T ) necessarily implies an Ω( √ Q T ) anytime regret on any sequence with quadratic variation Q T . The algorithms we consider are online linear optimization forecasters whose weight vector at time t + 1 is the gradient of a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2015
2015
2020
2020

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(7 citation statements)
references
References 11 publications
0
7
0
Order By: Relevance
“…We are not aware of any lower bounds in terms of the variance V k T . Gofer and Mansour [2012] provide lower bounds that hold for any sequence, in terms of the squared loss ranges in each round, but these do not apply to methods that adaptively tune their learning rate. For quantile bounds, Koolen [2013] takes a first step by characterizing the Pareto optimal quantile bounds for 2 experts in the √ T regime.…”
Section: Discussionmentioning
confidence: 99%
“…We are not aware of any lower bounds in terms of the variance V k T . Gofer and Mansour [2012] provide lower bounds that hold for any sequence, in terms of the squared loss ranges in each round, but these do not apply to methods that adaptively tune their learning rate. For quantile bounds, Koolen [2013] takes a first step by characterizing the Pareto optimal quantile bounds for 2 experts in the √ T regime.…”
Section: Discussionmentioning
confidence: 99%
“…For this notion, we show that, if all predictors are individually non-discriminatory with respect to equalized error rates, running separate multiplicative weights algorithms, one for each subpopulation, preserves this non-discrimination without decay in the efficiency (Theorem 3). The key property we use is that the multiplicative weights algorithm guarantees to perform not only no worse than the best predictor in hindsight but also no better; this property holds for a broader class of algorithms [GM16]. Our result applies to general loss functions beyond binary predictions and only requires predictors to satisfy the weakened assumption of being approximately non-discriminatory.…”
Section: Our Contributionmentioning
confidence: 91%
“…A crucial property we use is that multiplicative weights not only does not perform worse than the best expert; it also does not perform better. In fact, this property holds more generally by online learning algorithms with optimal regret guarantees [GM16].…”
Section: Fairness In Composition With Respect To An Alternative Metricmentioning
confidence: 93%
“…Algorithms exist that satisfy the even stronger property that the regret is at most ρ k and at least 0. Such non-negative-regret algorithms include all linear cost Regularized Follow the Leader algorithms, which includes Randomized Weighted Majority and linear cost Online Gradient Descent [22].…”
Section: Preliminariesmentioning
confidence: 99%