2019 International Conference on Robotics and Automation (ICRA) 2019
DOI: 10.1109/icra.2019.8794134
|View full text |Cite
|
Sign up to set email alerts
|

Crowd-Robot Interaction: Crowd-Aware Robot Navigation With Attention-Based Deep Reinforcement Learning

Abstract: This paper proposes an end-to-end deep reinforcement learning approach for mobile robot navigation with dynamic obstacles avoidance. Using experience collected in a simulation environment, a convolutional neural network (CNN) is trained to predict proper steering actions of a robot from its egocentric local occupancy maps, which accommodate various sensors and fusion algorithms. The trained neural network is then transferred and executed on a real-world mobile robot to guide its local path planning. The new ap… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
354
0
1

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 425 publications
(400 citation statements)
references
References 67 publications
(104 reference statements)
5
354
0
1
Order By: Relevance
“…There are other works that use Imitation Learning approaches to learn policies [16,18,23]. In the Crowd-Robot Interaction (CRI) approach [3], in which is mainly based the navigation in this paper, Imitation Learning is combined with the Deep Reinforcement Learning task to obtain a policy (SARL) for robot navigation. For Imitation Learning, the policy used to control all the agents is ORCA.…”
Section: Navigation Based On Deep Reinforcement Learningmentioning
confidence: 99%
See 2 more Smart Citations
“…There are other works that use Imitation Learning approaches to learn policies [16,18,23]. In the Crowd-Robot Interaction (CRI) approach [3], in which is mainly based the navigation in this paper, Imitation Learning is combined with the Deep Reinforcement Learning task to obtain a policy (SARL) for robot navigation. For Imitation Learning, the policy used to control all the agents is ORCA.…”
Section: Navigation Based On Deep Reinforcement Learningmentioning
confidence: 99%
“…The states for all the agents and obstacles are always known and used to calculate the value function using a neural network and the Temporal Difference algorithm of the CRI work [3]. With this value function the greedy policy is calculated, which gives in this case the robot's velocity(the optimal action) each timestep.…”
Section: Problem Formulationmentioning
confidence: 99%
See 1 more Smart Citation
“…However, the model parameters need to be tuned for different application scenarios. More recent research has used deep reinforcement learning (RL) successfully to learn efficient policies that model the cooperation and interactions implicitly [9]- [11]. We propose to represent the interactions among humans and robot as a graph with both nodes and edges varying over time.…”
Section: Introductionmentioning
confidence: 99%
“…Previously proposed approaches differ primarily in the structure of the networks used to encode the robot-crowd state and to estimate the corresponding value. There are several limitations of current models, which cause their performance to degrade when the crowd density increases [9], [10]. First, existing models have considered only pairwise interactions between the robot and each human in the crowd.…”
Section: Introductionmentioning
confidence: 99%