2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2020
DOI: 10.1109/iros45743.2020.9341620
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Exploration in Constrained Environments with Goal-Oriented Reference Path

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
15
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 22 publications
(15 citation statements)
references
References 24 publications
0
15
0
Order By: Relevance
“…For instance, [15] uses reinforcement learning (RL) to acquire local policy for steering within SMP methods. Similarly, [16] [17] constructs high-level landmarks, which are later connected by local RL policies. However, these methods inherit RL limitations such as requiring exhaustive interactions with their environments for learning.…”
Section: Related Workmentioning
confidence: 99%
“…For instance, [15] uses reinforcement learning (RL) to acquire local policy for steering within SMP methods. Similarly, [16] [17] constructs high-level landmarks, which are later connected by local RL policies. However, these methods inherit RL limitations such as requiring exhaustive interactions with their environments for learning.…”
Section: Related Workmentioning
confidence: 99%
“…It also performed in a scenario with multiple humans simulating a crowd. Ota et al [20] developed an approch based on the SAC technique to efficiently navigate. It was also done in simulation, with the agent learning the course.…”
Section: Related Workmentioning
confidence: 99%
“…In our proposed work, this is achieved as the path encodes the goal and automatically tries to avoid obstacles in the environment, too. The closest work to ours is [12,18]. [12] learns an RL agent that optimizes trajectory for a 6-DoF manipulator arm.…”
Section: Related Workmentioning
confidence: 99%
“…Our approach, however, can deal with changes in the environment by conditioning an RL agent with a reference path which is also generated based on an observation of the current environment. [18] considers a similar setting for 2-dimensional robot navigation tasks only in simulation. Our method, however, considers higher-dimensional path planning, and is evaluated on real systems as well as simulated ones.…”
Section: Related Workmentioning
confidence: 99%