2012 IEEE 27th Convention of Electrical and Electronics Engineers in Israel 2012
DOI: 10.1109/eeei.2012.6376912
|View full text |Cite
|
Sign up to set email alerts
|

Stochastic bandits with pathwise constraints

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2014
2014
2019
2019

Publication Types

Select...
1
1

Relationship

2
0

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 12 publications
0
2
0
Order By: Relevance
“…Multi-Armed Bandits (MABs) are a well-known framework in machine learning [7]. They succeed in capturing the trade-off between exploration and exploitation in sequential decision problems, and have been used in the context of learning in CRNs over the last few years [13], [5,6]. Classical bandit problems comprise an agent (user) repeatedly choosing a single option (arm) from a set of options whose characteristics are initially unknown, receiving a certain reward based on every choice.…”
Section: Multi-armed Banditsmentioning
confidence: 99%
“…Multi-Armed Bandits (MABs) are a well-known framework in machine learning [7]. They succeed in capturing the trade-off between exploration and exploitation in sequential decision problems, and have been used in the context of learning in CRNs over the last few years [13], [5,6]. Classical bandit problems comprise an agent (user) repeatedly choosing a single option (arm) from a set of options whose characteristics are initially unknown, receiving a certain reward based on every choice.…”
Section: Multi-armed Banditsmentioning
confidence: 99%
“…Using MABs to model CRNs was first suggested in [11], in a rather straightforward manner -the channels of a communication network simply correspond to the arms of a MAB. An extension that also takes operational constraints into account appears in [12].…”
Section: The Crn-mab Frameworkmentioning
confidence: 99%