2020 IEEE International Conference on Robotics and Automation (ICRA) 2020
DOI: 10.1109/icra40945.2020.9197209
|View full text |Cite
|
Sign up to set email alerts
|

Cooperative Multi-Robot Navigation in Dynamic Environment with Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 36 publications
(18 citation statements)
references
References 15 publications
0
14
0
Order By: Relevance
“…RL-NRVO is the policy using the original information of robots as input instead of RVO vectors. This information includes relative positions/velocities and robot radiuses, as introduced in [13], i.e., o sur = [p x , p y , v x , v y , R]. Similarly, RL-LSTM is the policy which replaces the BiGRUs with LSTM to tackle the input information of robots with varying numbers.…”
Section: B Results and Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…RL-NRVO is the policy using the original information of robots as input instead of RVO vectors. This information includes relative positions/velocities and robot radiuses, as introduced in [13], i.e., o sur = [p x , p y , v x , v y , R]. Similarly, RL-LSTM is the policy which replaces the BiGRUs with LSTM to tackle the input information of robots with varying numbers.…”
Section: B Results and Discussionmentioning
confidence: 99%
“…However, the dimension of the input data for a neural network is required to be fixed. Thus, for the environment model with the time-varying number of surrounding robots, some approaches assume that the number of obstacles is a constant and has an upper limit [13]. RNNs are able to tackle a variable number of moving obstacles, such as in GA3C-CADRL [11], where the exteroceptive measurements at each step are rearranged as the sequential input data through the long shortterm memory (LSTM) module to produce a fixed-size feature vector of the environment.…”
Section: B Deep Reinforcement Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…RL is being extensively used for the navigation of ground robots by mapping raw sensor measurements to its navigation commands for obstacle avoidance (Fan et al, 2020). For instance, Han et al (2020) trained a homogenous multi-agent system of navigating ground robots that used proximal policy optimization to maximize target and obstacle avoidance rewards based on their poses and velocities. This approach is suitable for the navigation control of mobile robots to various machines on a manufacturing shop floor.…”
Section: Iced21mentioning
confidence: 99%
“…Lin et al [10] aim to navigate the geometric center of a robot team to reach waypoints and develop an end-to-end policy shared among the robots that consider raw laser data inputs and position information of other robots. Han et al [22] also consider observable states of other robots and dynamic obstacles with Gaussian noises. They propose a greedy target allocation method to efficiently navigate the robots to multiple targets.…”
Section: Related Workmentioning
confidence: 99%