2019
DOI: 10.48550/arxiv.1906.01202
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning Transferable Cooperative Behavior in Multi-Agent Teams

Abstract: While multi-agent interactions can be naturally modeled as a graph, the environment has traditionally been considered as a black box. We propose to create a shared agent-entity graph, where agents and environmental entities form vertices, and edges exist between the vertices which can communicate with each other. Agents learn to cooperate by exchanging messages along the edges of this graph. Our proposed multi-agent reinforcement learning framework is invariant to the number of agents or entities present in th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
28
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(28 citation statements)
references
References 13 publications
(17 reference statements)
0
28
0
Order By: Relevance
“…Methods for Dynamic Team Compositions Several recent works in transfer learning and curriculum learning attempt to transfer policy of small teams for larger teams (Carion et al, 2019;Shu & Tian, 2019;Agarwal et al, 2019;Wang et al, 2020b;Long et al, 2020). These works mostly consider teams with different numbers of homogeneous agents.…”
Section: Centralized Training With Decentralized Executionmentioning
confidence: 99%
“…Methods for Dynamic Team Compositions Several recent works in transfer learning and curriculum learning attempt to transfer policy of small teams for larger teams (Carion et al, 2019;Shu & Tian, 2019;Agarwal et al, 2019;Wang et al, 2020b;Long et al, 2020). These works mostly consider teams with different numbers of homogeneous agents.…”
Section: Centralized Training With Decentralized Executionmentioning
confidence: 99%
“…Multi-Actor-Attention-Critic (MAAC) is proposed in [19] to aggregate information using attention mechanism from all the other agents. Similarly, [11], [13], [20] also employ the attention mechanism to learn a representation for the action-value function. However, the communication graphs used there are either dense or ad-hoc (k nearest neighbors), which makes the learning difficult.…”
Section: A Related Workmentioning
confidence: 99%
“…Graph neural network (GNN) is a specific neural network architecture in which permutation-invariance features can be embedded via graph pooling operations, so this approach has been applied in MARL [11]- [13] to exploit the interchangeability. As MARL is a non-structural scenario where the links/connections between the nodes/agents are ambiguous to decide, a graph has to be created in advance to apply GNN for MARL.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Existing CTDE research covers important topics such as division of agents [27], diversification [32] and exploration [19]. Recent works [29,11,1,17,16] have also started to make progress in transfer learning in cooperative MARL. For example, Liu et al [16] use policy distillation [22] to achieve fixed agent transfer learning.…”
Section: Introductionmentioning
confidence: 99%