2023
DOI: 10.1109/lra.2023.3246844
|View full text |Cite
|
Sign up to set email alerts
|

Overcoming Exploration: Deep Reinforcement Learning for Continuous Control in Cluttered Environments From Temporal Logic Specifications

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(8 citation statements)
references
References 25 publications
0
8
0
Order By: Relevance
“…This approach requires the ability to plan-ahead in the environment, which is not always feasible. Automaton-guided RL has been used to aid navigational exploration for robotic domains (Cai et al 2023) and for multi-agent settings (Hammond et al 2021). Generating a curriculum given the high-level objective (Shukla et al 2023) requires access to the Object-Oriented MDP (Diuk, Cohen, and Littman 2008), which cannot be obtained if environment details are not known in advance.…”
Section: Related Workmentioning
confidence: 99%
“…This approach requires the ability to plan-ahead in the environment, which is not always feasible. Automaton-guided RL has been used to aid navigational exploration for robotic domains (Cai et al 2023) and for multi-agent settings (Hammond et al 2021). Generating a curriculum given the high-level objective (Shukla et al 2023) requires access to the Object-Oriented MDP (Diuk, Cohen, and Littman 2008), which cannot be obtained if environment details are not known in advance.…”
Section: Related Workmentioning
confidence: 99%
“…q K+M P with q K P = q K+M P is executed infinitely often. Following prior work [6], we can decompose the optimal path τ F = τ pre P [τ suf P ] ω based on automaton components into a sequence of goal-reaching trajectories i.e., τ F = τ 0 τ 1 . .…”
Section: A Offline Motion Planningmentioning
confidence: 99%
“…Each τ i can be presented as an optimal solution of reachability navigation expressed as a simple LTL formula ϕ i,F = □¬O ∧ ϕ gi , O represent obstacles. We refer readers for more details about the decomposition procedure in [6].…”
Section: A Offline Motion Planningmentioning
confidence: 99%
“…A compositional RL algorithm that interleaves Dijkastra's algorithm for high-level task planning with learning sub-task policy using RL is developed in Jothimurugan et al (2021). Cai et al (2023) introduce a path planning-guided reward design scheme to RL-based policy design with LTL specified mission goals. Cai et al (2021) consider motion planning under LTL task specifications in continuous state and action spaces, and develop an unsupervised one-shot and on-the-fly motion planning framework to learn the unknown state.…”
Section: Continuous-state Pomdpmentioning
confidence: 99%