2017
DOI: 10.1007/978-3-319-63703-7_22
|View full text |Cite
|
Sign up to set email alerts
|

Bandit Models of Human Behavior: Reward Processing in Mental Disorders

Abstract: Drawing an inspiration from behavioral studies of human decision making, we propose here a general parametric framework for multi-armed bandit problem, which extends the standard Thompson Sampling approach to incorporate reward processing biases associated with several neurological and psychiatric conditions, including Parkinson's and Alzheimer's diseases, attention-deficit/hyperactivity disorder (ADHD), addiction, and chronic pain. We demonstrate empirically that the proposed parametric approach can often out… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
22
0
1

Year Published

2018
2018
2024
2024

Publication Types

Select...
3
2
2

Relationship

3
4

Authors

Journals

citations
Cited by 12 publications
(23 citation statements)
references
References 17 publications
0
22
0
1
Order By: Relevance
“…We now outlined the split models evaluated in our three settings: the MAB case with the Human-Based Thompson Sampling (HBTS) [10], the CB case with the Split Contextual Thompson Sampling (SCTS), and the RL case with the Split Q-Learning [29,32]. All three split agent classes are standardized for their parametric notions (see Table 1 for a complete parametrization and Appendix A for more literature review of these clinically-inspired reward-processing biases).…”
Section: Two-stream Split Models In Mab Cb and Rlmentioning
confidence: 99%
See 2 more Smart Citations
“…We now outlined the split models evaluated in our three settings: the MAB case with the Human-Based Thompson Sampling (HBTS) [10], the CB case with the Split Contextual Thompson Sampling (SCTS), and the RL case with the Split Q-Learning [29,32]. All three split agent classes are standardized for their parametric notions (see Table 1 for a complete parametrization and Appendix A for more literature review of these clinically-inspired reward-processing biases).…”
Section: Two-stream Split Models In Mab Cb and Rlmentioning
confidence: 99%
“…Split Multi-Armed Bandit Model. The split MAB agent is built upon Human-Based Thompson Sampling (HBTS, Algorithm 1) [10]. The positive and negative streams are each stored in the success and failure counts S a and F a .…”
Section: Two-stream Split Models In Mab Cb and Rlmentioning
confidence: 99%
See 1 more Smart Citation
“…Homeostatic decision-making has been combined with models for hormonal control (Avila-García and Cañamero, 2005) and cognitive modulation (Bach, 2015). It also supports decision-making influenced by individual personality traits (Bouneffouf, Rish, and Cecchi, 2017). For modeling animals with multiple needs, e.g.…”
Section: Animatsmentioning
confidence: 99%
“…to regulate their homeostatic variables and thus survive as long as possible (Keramati and Gutkin, 2011;Yoshida, 2017). Homeostatic decision-making combines naturally with models for hormonal control (Avila-García and Cañamero, 2005), cognitive modulation (Bach, 2015), and personality traits (Bouneffouf, Rish, and Cecchi, 2017). Moreover, homeostatic agents can be naturally linked to reinforcement learning by defining reward as the difference in need status from one time to another.…”
Section: Introductionmentioning
confidence: 99%