Proceedings of the 2017 ACM Conference on Economics and Computation 2017
DOI: 10.1145/3033274.3085145
|View full text |Cite
|
Sign up to set email alerts
|

Online Auctions and Multi-scale Online Learning

Abstract: We consider revenue maximization in online auctions and pricing. A seller sells an identical item in each period to a new buyer, or a new set of buyers. For the online posted pricing problem, we show regret bounds that scale with the best fixed price, rather than the range of the values. We also show regret bounds that are almost scale free, and match the offline sample complexity, when comparing to a benchmark that requires a lower bound on the market share. These results are obtained by generalizing the clas… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
45
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
6
1

Relationship

2
5

Authors

Journals

citations
Cited by 24 publications
(45 citation statements)
references
References 23 publications
0
45
0
Order By: Relevance
“…We present our result in a more general setting where the learner receives a predicted loss vector m t before deciding w t (Rakhlin and Sridharan, 2013b), and show a bound REG(e i ) = Õ (ln d) T t=1 (ℓ t,i − m t,i ) 2 simultaneously for all i (setting m t = 0 resolves the original impossible tuning issue). Using different m t , we achieve various regret bounds summarized in Table 1, which either recover the guarantees of existing algorithms such as (A, B)-PROD (Sani et al, 2014), ADAPT-ML-PROD (Gaillard et al, 2014), OPTIMISTIC-ADAPT-ML-PROD (Wei et al, 2016), or improve over existing variance/path-length/multi-scale bounds in (Steinhardt and Liang, 2014;Bubeck et al, 2017;Foster et al, 2017;Cutkosky and Orabona, 2018). Notably, we achieve a bound Õ (ln d) T t=1 w t − e i , ℓ t − ℓ t−1 2 which simultaneously ensures the "fast rate" consequences discussed in for stochastic settings and the path-length bound useful for fast convergence in games (Syrgkanis et al, 2015).…”
Section: Notesmentioning
confidence: 62%
See 2 more Smart Citations
“…We present our result in a more general setting where the learner receives a predicted loss vector m t before deciding w t (Rakhlin and Sridharan, 2013b), and show a bound REG(e i ) = Õ (ln d) T t=1 (ℓ t,i − m t,i ) 2 simultaneously for all i (setting m t = 0 resolves the original impossible tuning issue). Using different m t , we achieve various regret bounds summarized in Table 1, which either recover the guarantees of existing algorithms such as (A, B)-PROD (Sani et al, 2014), ADAPT-ML-PROD (Gaillard et al, 2014), OPTIMISTIC-ADAPT-ML-PROD (Wei et al, 2016), or improve over existing variance/path-length/multi-scale bounds in (Steinhardt and Liang, 2014;Bubeck et al, 2017;Foster et al, 2017;Cutkosky and Orabona, 2018). Notably, we achieve a bound Õ (ln d) T t=1 w t − e i , ℓ t − ℓ t−1 2 which simultaneously ensures the "fast rate" consequences discussed in for stochastic settings and the path-length bound useful for fast convergence in games (Syrgkanis et al, 2015).…”
Section: Notesmentioning
confidence: 62%
“…Our first main contribution is to show that, perhaps surprisingly, this impossible tuning is in fact possible (up to an additional ln T factor), via an algorithm combining ideas that mostly appear before already. More concretely, we achieve this via Mirror Descent with a correction term similar to (Steinhardt and Liang, 2014) and a weighted negative entropy regularizer with different learning rates for each expert (and each round) similar to (Bubeck et al, 2017). Note that while natural, this algorithm has not been studied before, 1 and is not equivalent to using different learning rates for different experts in PROD or multiplicative-weight, as it does not admit a closed "proportional" form (and instead needs to be computed via a line search).…”
Section: Notesmentioning
confidence: 99%
See 1 more Smart Citation
“…where H t is a multiset of learning rates we define below. Similar regularizers have been used by Bubeck et al (2017) and Chen et al (2021) to obtain multiscale algorithms. In particular, the algorithm of Chen et al ( 2021) is related to ours, as they also use corrections.…”
Section: Full Information Settingmentioning
confidence: 93%
“…input sequence. In particular this is true for a monopolistic seller learning an optimal price or an optimal auction Bubeck et al (2017); Blum and Hartline (2005). It is tempting to conjecture that the same holds for our setting as well, but we run into difficulties even modeling the problem.…”
Section: Future Workmentioning
confidence: 99%