2017
DOI: 10.1098/rsos.171377
|View full text |Cite
|
Sign up to set email alerts
|

Risk-aware multi-armed bandit problem with application to portfolio selection

Abstract: Sequential portfolio selection has attracted increasing interest in the machine learning and quantitative finance communities in recent years. As a mathematical framework for reinforcement learning policies, the stochastic multi-armed bandit problem addresses the primary difficulty in sequential decision-making under uncertainty, namely the exploration versus exploitation dilemma, and therefore provides a natural connection to portfolio selection. In this paper, we incorporate risk awareness into the classic m… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
20
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
9
1

Relationship

1
9

Authors

Journals

citations
Cited by 49 publications
(27 citation statements)
references
References 49 publications
0
20
0
1
Order By: Relevance
“…They are commonly known to have applications in medical trials (Armitage [4] or Anscombe [3]) and experimental design (Berry and Fristedt [13] or the classic paper of Robbins [69]), along with other areas. A few recent works in finance for portfolio selection can also be found in Huo and Fu [45] or Shen et al [72]. The basic idea is that one has M 'bandits 2 ', or equivalently, a bandit with M arms, and one must choose which bandit should be played at each time.…”
Section: Multi-armed Banditsmentioning
confidence: 99%
“…They are commonly known to have applications in medical trials (Armitage [4] or Anscombe [3]) and experimental design (Berry and Fristedt [13] or the classic paper of Robbins [69]), along with other areas. A few recent works in finance for portfolio selection can also be found in Huo and Fu [45] or Shen et al [72]. The basic idea is that one has M 'bandits 2 ', or equivalently, a bandit with M arms, and one must choose which bandit should be played at each time.…”
Section: Multi-armed Banditsmentioning
confidence: 99%
“…Of particular interest is to find the existence of gamblers who are clearly loss-exiting against gain-exiting. Besides, understanding individual gambling behaviour of choosing between and deploying betting systems that have varying risks from the perspective of the multi-armed bandit problem [ 19 ] is both interesting and promising.…”
Section: Discussionmentioning
confidence: 99%
“…Risk-aware methods [8,17,41] are online learning algorithms that model the risk associated with executing certain actions as a cost which is to be constrained and minimized. Galichet et al [8] introduce the concept of risk-awareness for the multi-armed bandit framework.…”
Section: Safety In Contextual Banditsmentioning
confidence: 99%