Robotics: Science and Systems XVI 2020
DOI: 10.15607/rss.2020.xvi.097
|View full text |Cite
|
Sign up to set email alerts
|

Explaining Multi-stage Tasks by Learning Temporal Logic Formulas from Suboptimal Demonstrations

Abstract: We present a method for learning to perform multistage tasks from demonstrations by learning the logical structure and atomic propositions of a consistent linear temporal logic (LTL) formula. The learner is given successful but potentially suboptimal demonstrations, where the demonstrator is optimizing a cost function while satisfying the LTL formula, and the cost function is uncertain to the learner. Our algorithm uses the Karush-Kuhn-Tucker (KKT) optimality conditions of the demonstrations together with a co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
19
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 20 publications
(19 citation statements)
references
References 32 publications
0
19
0
Order By: Relevance
“…logic or automata, which admit well defined compositions and explicitly encode temporal constraints. The development of this literature mirrors the historical path taken in reward based research, with works adapting optimal control [7,2], Bayesian [11,17], and maximum entropy approaches [14,15] IRL approaches.…”
Section: Related Workmentioning
confidence: 96%
See 2 more Smart Citations
“…logic or automata, which admit well defined compositions and explicitly encode temporal constraints. The development of this literature mirrors the historical path taken in reward based research, with works adapting optimal control [7,2], Bayesian [11,17], and maximum entropy approaches [14,15] IRL approaches.…”
Section: Related Workmentioning
confidence: 96%
“…In particular, and in contrast to the reward setting, the discrete nature of automata and logic, combined with the assumed a-priori ignorance of the relevant memory required to describe the task, makes existing gradient based approaches either intractable or inapplicable. Instead, current literature either enumerates concepts [14,2,11,17] or hill climbs via simple probabilistic mutations [7,1].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…They search for automata using DPLL search. [10] learns an LTL formula and proposition mapping from demonstrations. Their approach relies on counterexample generation and testing.…”
Section: Related Workmentioning
confidence: 99%
“…Applications include program specification [20], anomaly and fault detection [4], robotics [6], and many more: we refer to [5], Section 7, for a list of practical applications. An equivalent point of view on LTL learning is as a specification mining question.…”
Section: State Of the Artmentioning
confidence: 99%