2022
DOI: 10.1177/02783649221082115
|View full text |Cite
|
Sign up to set email alerts
|

Backpropagation through signal temporal logic specifications: Infusing logical structure into gradient-based methods

Abstract: This paper presents a technique, named STLCG, to compute the quantitative semantics of Signal Temporal Logic (STL) formulas using computation graphs. STLCG provides a platform which enables the incorporation of logical specifications into robotics problems that benefit from gradient-based solutions. Specifically, STL is a powerful and expressive formal language that can specify spatial and temporal properties of signals generated by both continuous and hybrid systems. The quantitative semantics of STL provide … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
20
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(20 citation statements)
references
References 26 publications
0
20
0
Order By: Relevance
“…We briefly summarize Eq. ( 1) here, but refer readers to Leung et al [17] (Section 2.2) for a pedagogical introduction to STL. The core of an STL formula are predicates µ c of the form µ(z) > c, where c ∈ R and µ : R n → R is a differentiable function.…”
Section: Signal Temporal Logic (Stl)mentioning
confidence: 99%
See 2 more Smart Citations
“…We briefly summarize Eq. ( 1) here, but refer readers to Leung et al [17] (Section 2.2) for a pedagogical introduction to STL. The core of an STL formula are predicates µ c of the form µ(z) > c, where c ∈ R and µ : R n → R is a differentiable function.…”
Section: Signal Temporal Logic (Stl)mentioning
confidence: 99%
“…By construction, every STL formula admits a robustness formula measuring the degree of rule satisfaction, which is used as the guide J . Since it is necessary to compute a gradient through this function, STL formulas are implemented using differentiable frameworks [17], [18]. Multi-agent guidance.…”
Section: : While Not Done Domentioning
confidence: 99%
See 1 more Smart Citation
“…Since the QP ( 9) is differentiable with respect to its parameters using the technique in [25], we backpropagate the gradient of the objective funtion in (19) through the QP to all parameters θ. The gradients of the STL robustness are calculated analytically and automatically using an adapted version of STLCG [26] that use the robustness in [23]. Then we update the parameters using the gradient.…”
Section: Learning Robust Controllersmentioning
confidence: 99%
“…In this setting, logical specifications serve primarily to help generate the reward function used by a DRL procedure; this approach is known as reward shaping. However, as we show in this paper, by equipping these enriched semantics with differentiable operators [13], [14], policy updates can be meaningfully constrained to yield a significantly more sample-efficient learning technique compared with existing reward-shaping methods.…”
Section: Introductionmentioning
confidence: 99%