2017 IEEE International Conference on Robotics and Automation (ICRA) 2017
DOI: 10.1109/icra.2017.7989381
|View full text |Cite
|
Sign up to set email alerts
|

Target-driven visual navigation in indoor scenes using deep reinforcement learning

Abstract: Two less addressed issues of deep reinforcement learning are (1) lack of generalization capability to new target goals, and (2) data inefficiency i.e., the model requires several (and often costly) episodes of trial and error to converge, which makes it impractical to be applied to real-world scenarios. In this paper, we address these two issues and apply our model to the task of target-driven visual navigation. To address the first issue, we propose an actor-critic model whose policy is a function of the goal… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

3
1,060
0
4

Year Published

2017
2017
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 1,288 publications
(1,116 citation statements)
references
References 41 publications
3
1,060
0
4
Order By: Relevance
“…Another promising approach has been use simulations to get training data for reinforcement or imitation learning tasks, while testing the learned policy in the real world [14], [15], [11]. Clearly, this approach suffers from the domain shift between simulation and reality and might require some realworld data to be able to generalize [11].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Another promising approach has been use simulations to get training data for reinforcement or imitation learning tasks, while testing the learned policy in the real world [14], [15], [11]. Clearly, this approach suffers from the domain shift between simulation and reality and might require some realworld data to be able to generalize [11].…”
Section: Related Workmentioning
confidence: 99%
“…These methodologies can be divided into two main categories: (i) methods based on reinforcement learning (RL) [7], [11] and (ii) methods based on supervised learning [6], [12], [9], [10], [13].…”
Section: Related Workmentioning
confidence: 99%
“…Asynchronous parallel sampling is applied to a number of deep reinforcement learning algorithms including DQN and an actor-critic method (called A3C) which achieved state of the art results on the Atari benchmark and demonstrated effective performance in a navigation task in a 3D simulated environment. Utilizing additional information about the target is used to speed up and improve deep RL in [11]. Information about the target is provided in the form of images of the target.…”
Section: Related Workmentioning
confidence: 99%
“…For comparison, in our experiments on a 30x30 Grid, the shortest route to the target consists of 57 actions. While a recent study proposed a new complex simulator with realistic graphics [11] where the shortest trajectory to the target is typically less than 20 actions. …”
Section: A Grid Navigation Taskmentioning
confidence: 99%
See 1 more Smart Citation