2020
DOI: 10.48550/arxiv.2002.10525
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Scalable Multi-Agent Inverse Reinforcement Learning via Actor-Attention-Critic

Abstract: Multi-agent adversarial inverse reinforcement learning (MA-AIRL) is a recent approach that applies single-agent AIRL to multi-agent problems where we seek to recover both policies for our agents and reward functions that promote expertlike behavior. While MA-AIRL has promising results on cooperative and competitive tasks, it is sample-inefficient and has only been validated empirically for small numbers of agents -its ability to scale to many agents remains an open question. We propose a multi-agent inverse RL… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 6 publications
0
1
0
Order By: Relevance
“…Each agent receives a state input s i , covering the global s and a specification of the agent's location, and generates its policy π i = π(s i ) and state value V i = V (s i ). The model is shared between agents, adding scalability when the number of agents grows larger and the environment more complex (Iqbal and Sha, 2018;Jiang and Lu, 2018;Jeon et al, 2020). It also potentially alleviates instability due to the non-stationary nature of multi-agent environments by sharing the same embedding space (Lowe et al, 2017).…”
Section: Multiagent Reinforcement Learningmentioning
confidence: 99%
“…Each agent receives a state input s i , covering the global s and a specification of the agent's location, and generates its policy π i = π(s i ) and state value V i = V (s i ). The model is shared between agents, adding scalability when the number of agents grows larger and the environment more complex (Iqbal and Sha, 2018;Jiang and Lu, 2018;Jeon et al, 2020). It also potentially alleviates instability due to the non-stationary nature of multi-agent environments by sharing the same embedding space (Lowe et al, 2017).…”
Section: Multiagent Reinforcement Learningmentioning
confidence: 99%