2007
DOI: 10.1063/1.2709596
|View full text |Cite
|
Sign up to set email alerts
|

An introduction to stochastic control theory, path integrals and reinforcement learning

Abstract: Abstract. Control theory is a mathematical description of how to act optimally to gain future rewards. In this paper I give an introduction to deterministic and stochastic control theory and I give an overview of the possible application of control theory to the modeling of animal behavior and learning. I discuss a class of non-linear stochastic control problems that can be efficiently solved using a path integral or by MC sampling. In this control formalism the central concept of cost-to-go becomes a free ene… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
116
0

Year Published

2008
2008
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 96 publications
(117 citation statements)
references
References 25 publications
(26 reference statements)
1
116
0
Order By: Relevance
“…However, one could argue that a sampling procedure to compute the path integral corresponds to a learning of the environment. A discussion on this line of thought can be found in (Kappen, 2007).…”
Section: Discussionmentioning
confidence: 99%
“…However, one could argue that a sampling procedure to compute the path integral corresponds to a learning of the environment. A discussion on this line of thought can be found in (Kappen, 2007).…”
Section: Discussionmentioning
confidence: 99%
“…In Kappen [7], control problems are associated with animal behavior. Living organisms, including human beings, employ cognitive resources to take decisions.…”
Section: Some Specific Applications and Contributionsmentioning
confidence: 99%
“…The fundamental point, though, is that one is dealing with the future; the plan is set now, at t=0, for an horizon that starts at t=0 and extends to a pre-defined future date. Therefore, as emphasized by Kappen [7], the control problem is stochastic in nature. There is uncertainty associated with future outcomes and the best the agent can do is to compute the optimal trajectory of some control variable(s) contingent on how the system is supposed to evolve.…”
Section: Why Stochastic Controls?mentioning
confidence: 99%
“…Optimal control [1], [2] aims at finding optimal behavior given a cost function and the dynamics of the system. Typically, the cost function consists of several objectives that need to be traded off by the experimenter.…”
Section: Introductionmentioning
confidence: 99%