1952
DOI: 10.1090/s0002-9904-1952-09620-8
|View full text |Cite
|
Sign up to set email alerts
|

Some aspects of the sequential design of experiments

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
380
0
6

Year Published

2004
2004
2018
2018

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 1,489 publications
(415 citation statements)
references
References 5 publications
1
380
0
6
Order By: Relevance
“…In this paper we analyze the Win-Stay, Lose-Switch (WSLS) model, also known as WinStay, Lose-Shift, which is used in psychology, game theory, statistics and machine learning [14], [15].…”
Section: Decision-making Modelmentioning
confidence: 99%
“…In this paper we analyze the Win-Stay, Lose-Switch (WSLS) model, also known as WinStay, Lose-Shift, which is used in psychology, game theory, statistics and machine learning [14], [15].…”
Section: Decision-making Modelmentioning
confidence: 99%
“…The well-studied multiarmed bandit problem was originally proposed by Robbins [7] in 1985. A gambler, firstly, chooses K slot machines to play.…”
Section: Multi-armed Bandit Problemmentioning
confidence: 99%
“…Additionally, among all the reinforcement learning techniques, a set of so-called multi-armed bandit algorithms is particularly suitable for the optimization of the network. That is, the number of transmissions in each sensor node can be furthermore modeled as a multiarmed bandit problem, originally described by Robins [7]. A multi-armed bandit, also called K-armed bandit, is similar to a traditional slot machine but generally with more than one lever.…”
Section: Introductionmentioning
confidence: 99%
“…The problem has its roots in the seminal works of Robbins (1952) and Bradt et al (1956), who focused on the much-studied case where engaging a project corresponds to sampling from a Bernoulli population with unknown success probability, the goal being to maximize the expected number of successes over T plays. An MDP formulation is obtained by a Bayesian approach, where a project/population state is its posterior distribution.…”
Section: Finite-horizon Multiarmed Banditsmentioning
confidence: 99%