2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2018
DOI: 10.1109/iros.2018.8593702
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning to Acquire Navigation Skills for Wheel-Legged Robots in Complex Environments

Abstract: Mobile robot navigation in complex and dynamic environments is a challenging but important problem. Reinforcement learning approaches fail to solve these tasks efficiently due to reward sparsities, temporal complexities and high-dimensionality of sensorimotor spaces which are inherent in such problems. We present a novel approach to train action policies to acquire navigation skills for wheel-legged robots using deep reinforcement learning. The policy maps heightmap image observations to motor commands to navi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
30
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 39 publications
(30 citation statements)
references
References 38 publications
0
30
0
Order By: Relevance
“…Chen et al [12] deployed also PPO for deep RL as we do. The authors rely on height-map observations as state representation for a wheel-legged robot.…”
Section: Related Workmentioning
confidence: 98%
“…Chen et al [12] deployed also PPO for deep RL as we do. The authors rely on height-map observations as state representation for a wheel-legged robot.…”
Section: Related Workmentioning
confidence: 98%
“…Kalashnikov et al used a scalable RL framework for learning vision-based dynamic grasping skills [17]. However, it is required lots of exploration to obtain a skill via an RL manner integrated with visuomotor framework, which is regarded to be data-inefficient [18,19].…”
Section: Related Workmentioning
confidence: 99%
“…However, our goal is to find a policy which is independent on the strip properties. One possibility of finding such a policy is to use domain randomization [18], [19]. Domain randomization for our task may be achieved by modifying the optimization (2) into:…”
Section: B Domain Randomizationmentioning
confidence: 99%