2008
DOI: 10.3758/cabn.8.4.429
|View full text |Cite
|
Sign up to set email alerts
|

Decision theory, reinforcement learning, and the brain

Abstract: Decision making is a core competence for animals and humans acting and surviving in environments they only partially comprehend, gaining rewards and punishments for their troubles. Decision-theoretic concepts permeate experiments and computational models in ethology, psychology, and neuroscience. Here, we review permeate experiments and computational models in ethology, psychology, and neuroscience. Here, we review a welln known, coherent Bayesian approach to decision making, showing how it unifies issues in M… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

10
486
1

Year Published

2010
2010
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 501 publications
(500 citation statements)
references
References 102 publications
10
486
1
Order By: Relevance
“…This features set could correspond to neural correlates of decision-making (23,24), probabilistic inference (13,25), and learning and strategy shifts (26,27) that were observed in single-unit recordings in primate and mammalian cortex. Seeking neural correlates of the model presented here would be of particular interest in light of the characterization of the role of memory systems involved in WP (28,29) and other learning and decision-making tasks (30,31), and theoretical models of incremental learning through spike timing-dependent plasticity (32,33).…”
Section: Discussionmentioning
confidence: 99%
“…This features set could correspond to neural correlates of decision-making (23,24), probabilistic inference (13,25), and learning and strategy shifts (26,27) that were observed in single-unit recordings in primate and mammalian cortex. Seeking neural correlates of the model presented here would be of particular interest in light of the characterization of the role of memory systems involved in WP (28,29) and other learning and decision-making tasks (30,31), and theoretical models of incremental learning through spike timing-dependent plasticity (32,33).…”
Section: Discussionmentioning
confidence: 99%
“…We make the common assumption that critic values are represented in ventral striatum and that phasic signals of dopamine convey the critic prediction error (Dayan & Daw, 2008;Montague et al, 1996;Roesch, Calu, & Schoenbaum, 2007). Actor learning.…”
Section: Opal Model Descriptionmentioning
confidence: 99%
“…(4) A model of the environment which includes a representation of the environment dynamics required for maximizing the sum of future rewards. These general RL concepts are further explained and elaborated by various authors (Dayan and Daw 2008;Kaelbling et al 1996;Montague et al 2004a;Sutton and Barto 1998;Worgotter and Porr 2005;Dayan and Abbott 2001).…”
Section: Introductionmentioning
confidence: 99%