2022
DOI: 10.1007/978-3-030-94583-1_7
|View full text |Cite
|
Sign up to set email alerts
|

Gradient-Descent for Randomized Controllers Under Partial Observability

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1
1

Relationship

3
4

Authors

Journals

citations
Cited by 10 publications
(11 citation statements)
references
References 43 publications
0
10
0
Order By: Relevance
“…Data availability. The tools used and data generated in our experimental evaluation are archived at DOI 10.5281/5568910 [26].…”
Section: Discussionmentioning
confidence: 99%
“…Data availability. The tools used and data generated in our experimental evaluation are archived at DOI 10.5281/5568910 [26].…”
Section: Discussionmentioning
confidence: 99%
“…A variety of approaches have been investigated including a branch-andbound algorithm [14] or mixed-integer linear programming (MILP) formulation [15,16]. Alternatively, one may search for randomised controllers via gradient descent [17] or via convex optimization [18Ű20]. Randomized controllers can be also extracted via deep reinforcement learning [21].…”
Section: Belief Exploration In Stormmentioning
confidence: 99%
“…The concrete start beliefs B start are determined via method select-beliefs (discussed below). Then, for each element, we execute an inductive search (l. [6][7][8][9][10][11][12][13][14][15][16][17][18] for the time given by t I /size(B set ), meaning we split the time for inductive search uniformly among all selected beliefs. We always produce an FSC from the initial belief and for this belief, we use the knowledge on how to treat memory updates (l. 13) and action prioritization (l. 14) in the same way as in Alg.…”
Section: Algorithm Overviewmentioning
confidence: 99%
See 1 more Smart Citation
“…Parameter synthesis is to find the right values for the unknown parameters with respect to a given constraint. Various synthesis techniques have been developed for parametric Markov chains (pMCs) ranging over e.g., the gradient-based methods [Heck et al, 2022], convex optimization [Cubuktepe et al, 2018;Cubuktepe et al, 2022], and region verification [Quatmann et al, 2016]. Recently, Salmani and Katoen [2021a] have proposed a translation from pBNs to pMCs that facilitates using pMC algorithms to analyze pBNs.…”
mentioning
confidence: 99%