2015 American Control Conference (ACC) 2015
DOI: 10.1109/acc.2015.7171174
|View full text |Cite
|
Sign up to set email alerts
|

Counterexample-guided permissive supervisor synthesis for probabilistic systems through learning

Abstract: Formal methods in robotic motion planning have emerged as a hot research topic recently due to its correctby-design nature, and most results haven been based on nonprobabilistic discrete models. To better handle the environment uncertainties, sensor noise and actuator imperfection, control problems in probabilistic systems like Markov Chain (MC) and Markov Decision Process (MDP) have also been studied. Most existing methods are either based on probabilistic model checking or through reinforcement learning orie… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
4
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
4
3
1

Relationship

3
5

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 19 publications
(18 reference statements)
1
4
0
Order By: Relevance
“…This paper merges and further develops our previous results [22], [23]. While the same L* learning algorithm [14] was considered in synthesis process, we changed the form of the supervisor and considered positive counterexamples so that the results can become less conservative.…”
Section: Introductionsupporting
confidence: 60%
See 2 more Smart Citations
“…This paper merges and further develops our previous results [22], [23]. While the same L* learning algorithm [14] was considered in synthesis process, we changed the form of the supervisor and considered positive counterexamples so that the results can become less conservative.…”
Section: Introductionsupporting
confidence: 60%
“…The resulting π i will then be passed to the third stage to answer the conjecture in our modified L* learning algorithm. Note that in our previous work [22], [23], we proposed to apply the hop-constrained k shortest path (HKSP) algorithm to find the smallest k paths so that eliminating them will make sure the DTMC will not be counterexample any more. The complexity of HKSP is higher than the HSP since HKSP has to search for k as well.…”
Section: B Single Agent Casementioning
confidence: 99%
See 1 more Smart Citation
“…The L* learning approach assumes the existence of a teacher who can answer the membership and equivalence queries (Angluin 1987;Wu and Lin 2015;Wu, Zhang, and Lin 2018;Zhang, Wu, and Lin 2015;. Our approach fulfills the role of the teacher by an RL engine and the queries are answered through interaction with the environment through the RL engine.…”
Section: Related Workmentioning
confidence: 99%
“…Also due to the partial observability in POMDPs, applying these methods to POMDP are fundamentally difficult. Besides these works, using the L * algorithm to learn system supervisor has also been considered in our previous work for MDPs [42]. To apply the L * algorithm to POMDP supervisor synthesis, in this paper, we extensively discuss the supervisor synthesis framework and design new membership query and conjecture checking rules to overcome the difficulties brought by partial observability.…”
Section: A Related Workmentioning
confidence: 99%