2020
DOI: 10.1609/icaps.v30i1.6751
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement Learning for Zone Based Multiagent Pathfinding under Uncertainty

Abstract: We address the problem of multiple agents finding their paths from respective sources to destination nodes in a graph (also called MAPF). Most existing approaches assume that all agents move at fixed speed, and that a single node accommodates only a single agent. Motivated by the emerging applications of autonomous vehicles such as drone traffic management, we present zone-based path finding (or ZBPF) where agents move among zones, and agents' movements require uncertain travel time. Furthermore, each zone can… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(10 citation statements)
references
References 15 publications
0
10
0
Order By: Relevance
“…MARL for MAPF: MAPF can be mapped to a Dec-POMDP instance in multiple ways to address different variants (Ma, Kumar, and Koenig 2017;Sartoretti et al 2019;Ling, Gupta, and Kumar 2020). We present the MAPF problem under uncertainty and partial observability using minimal assumptions to ensure the generality of our knowledge compilation framework.…”
Section: The Dec-pomdp Model and Mapfmentioning
confidence: 99%
See 4 more Smart Citations
“…MARL for MAPF: MAPF can be mapped to a Dec-POMDP instance in multiple ways to address different variants (Ma, Kumar, and Koenig 2017;Sartoretti et al 2019;Ling, Gupta, and Kumar 2020). We present the MAPF problem under uncertainty and partial observability using minimal assumptions to ensure the generality of our knowledge compilation framework.…”
Section: The Dec-pomdp Model and Mapfmentioning
confidence: 99%
“…We have integrated our knowledge compilation framework with two policy gradient approaches proposed in (Sartoretti et al 2019;Ling, Gupta, and Kumar 2020) (one using feedforward neural net, another using recurrent neural network based policy), and a QMIX-variant (Fu et al 2019) for MAPF, demonstrating the generalization power of the framework for a range of MARL solution methods.…”
Section: Policy Gradient Based Marlmentioning
confidence: 99%
See 3 more Smart Citations