2022
DOI: 10.48550/arxiv.2206.09546
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Policy Optimization with Linear Temporal Logic Constraints

Abstract: We study the problem of policy optimization (PO) with linear temporal logic (LTL) constraints. The language of LTL allows flexible description of tasks that may be unnatural to encode as a scalar cost function. We consider LTL-constrained PO as a systematic framework, decoupling task specification from policy selection, and an alternative to the standard of cost shaping. With access to a generative model, we develop a model-based approach that enjoys a sample complexity analysis for guaranteeing both task sati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 26 publications
0
2
0
Order By: Relevance
“…However, this prior work has not considered how to synergistically take advantage of policies with differentiable [13], [14] and programmable [30], [31], [32] formal constraints whose definitions can directly encode the high-level tasks. More recently, a constrained reinforcement learning framework with formal constraints [11] was proposed and provided a theoretical foundation for constrained reinforcement learning under formal constraints in a discrete state and action space. However, this work does not consider directly backpropagating through formal constraints and requires learning high-level abstractions, making it perform modestly in the high-dimensional continuous control tasks we consider.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…However, this prior work has not considered how to synergistically take advantage of policies with differentiable [13], [14] and programmable [30], [31], [32] formal constraints whose definitions can directly encode the high-level tasks. More recently, a constrained reinforcement learning framework with formal constraints [11] was proposed and provided a theoretical foundation for constrained reinforcement learning under formal constraints in a discrete state and action space. However, this work does not consider directly backpropagating through formal constraints and requires learning high-level abstractions, making it perform modestly in the high-dimensional continuous control tasks we consider.…”
Section: Related Workmentioning
confidence: 99%
“…Logic-Constrained Reinforcement Learning (LCRL) [1], [2], [3], [4], [5], [6], [7], [8], [9], [10], [11] is a rapidly growing field that enables end-to-end reinforcement learning to perform various high-level tasks. For example, a logical task constraint can require a drone to hover near a doorway around until the door is opened, execute a collection of complex instructions sequentially, or make decisions that are tailored to different conditions.…”
Section: Introductionmentioning
confidence: 99%