2013
DOI: 10.1371/journal.pcbi.1003150
|View full text |Cite|
|
Sign up to set email alerts
|

A Mixture of Delta-Rules Approximation to Bayesian Inference in Change-Point Problems

Abstract: Error-driven learning rules have received considerable attention because of their close relationships to both optimal theory and neurobiological mechanisms. However, basic forms of these rules are effective under only a restricted set of conditions in which the environment is stable. Recent studies have defined optimal solutions to learning problems in more general, potentially unstable, environments, but the relevance of these complex mathematical solutions to how the brain solves these problems remains uncle… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

7
127
1

Year Published

2014
2014
2024
2024

Publication Types

Select...
8
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 93 publications
(135 citation statements)
references
References 39 publications
(39 reference statements)
7
127
1
Order By: Relevance
“…Thus, nominally, the reward prediction errors during adaptation would be negative (i.e., reward is less than expected), because performance is worse under the perturbation compared with at baseline. Whether subjects interpret a given action as an improvement (a positive reward prediction error) or continued failure (a negative reward prediction error) may depend on whether they detect that a change point had occurred in the experiment following the rotation onset (Wilson et al 2013). If the imposed rotation is interpreted as a change, actions and/or strategies that reduce the initially large errors experienced after the onset of the perturbation may be associated with a positive reward prediction error and thus may be remembered.…”
Section: Discussionmentioning
confidence: 99%
“…Thus, nominally, the reward prediction errors during adaptation would be negative (i.e., reward is less than expected), because performance is worse under the perturbation compared with at baseline. Whether subjects interpret a given action as an improvement (a positive reward prediction error) or continued failure (a negative reward prediction error) may depend on whether they detect that a change point had occurred in the experiment following the rotation onset (Wilson et al 2013). If the imposed rotation is interpreted as a change, actions and/or strategies that reduce the initially large errors experienced after the onset of the perturbation may be associated with a positive reward prediction error and thus may be remembered.…”
Section: Discussionmentioning
confidence: 99%
“…Hierarchical Bayesian models have proven powerful for explaining the adaptation of behaviour to probabilistic contexts in volatile environments [1,15,17,70,7274,84,104,105]. Here we developed a novel instantiation of the HGF model [9] with a focus on TMs and two components: a three-level perceptual model and a response model (Fig 3A).…”
Section: Methodsmentioning
confidence: 99%
“…This set includes tasks based on the basic problems of foraging theory, including the patch-leaving problem, the diet selection problem, the central place foraging problem, and so forth (Stephens & Krebs, 1986). It also includes stopping problems and other classic optimization problems, such as the k-arm bandit problems, horizon problems, and change point detection problems (Pearson, Hayden, Raghavachari, & Platt, 2009;Wilson, Geana, White, Ludvig, & Cohen, 2014;Wilson, Nassar, & Gold, 2013). Indeed, it may also include variants of the intertemporal choice task in which the postreward delays are clearly cued (Pearson et al, 2010).…”
Section: Suggestions For Future Researchmentioning
confidence: 99%