2020
DOI: 10.1007/978-3-030-63461-2_1
|View full text |Cite
|
Sign up to set email alerts
|

Formal Policy Synthesis for Continuous-State Systems via Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
3

Relationship

2
7

Authors

Journals

citations
Cited by 18 publications
(10 citation statements)
references
References 25 publications
0
7
0
Order By: Relevance
“…The work provides control strategies maximizing the probability of satisfaction over unknown continuous-space dt-SCS while providing probabilistic closeness guarantees in the form of (2.3). Similarly based on the scheme in Figure 13, extensions to continuous spaces and ω-regular properties with formal guarantees are studied in [HAK19b,HKA20,KS20]. This line of works leads to the following problem.…”
Section: Temporal Logic Verification and Synthesismentioning
confidence: 99%
See 1 more Smart Citation
“…The work provides control strategies maximizing the probability of satisfaction over unknown continuous-space dt-SCS while providing probabilistic closeness guarantees in the form of (2.3). Similarly based on the scheme in Figure 13, extensions to continuous spaces and ω-regular properties with formal guarantees are studied in [HAK19b,HKA20,KS20]. This line of works leads to the following problem.…”
Section: Temporal Logic Verification and Synthesismentioning
confidence: 99%
“…A data-driven technique for satisfying temporal properties on unknown stochastic processes with continuous spaces is recently presented in [KS20]. The proposed framework is based on reinforcement learning that is used to compute sub-optimal policies that are finite-memory and deterministic.…”
Section: Directions For Open Researchmentioning
confidence: 99%
“…When complex logical properties are of interest, e.g., those expressed as linear temporal logic formulae over finite traces (a.k.a. LTL F formulae [14]), results in [15,16,17] provide formal safety guarantees for AI-based controllers by considering the desired properties in the reward functions. Note that these results are only applicable to those AI-based controllers whose reward functions are easy to be designed, while reward functions for some control tasks are difficult to be obtained (e.g., [18]).…”
Section: Introductionmentioning
confidence: 99%
“…The closest line of work to ours, which aims to avoid HRL requirements, are model-based (Fu and Topcu 2014;Sadigh et al 2014;Fulton and Platzer 2018;Cai et al 2021) or model-free RL approaches that constrain the agent with a temporal logic property (Hasanbeig et al 2018;Toro Icarte et al 2018;Camacho et al 2019;Hasanbeig et al 2019a;Yuan et al 2019;De Giacomo et al 2019, 2020Hasanbeig et al 2019dHasanbeig et al ,c, 2020bKazemi and Soudjani 2020;Lavaei et al 2020). These approaches are limited to finitestate systems, or more importantly require the temporal logic formula to be known a priori.…”
Section: Introductionmentioning
confidence: 99%