2021
DOI: 10.1609/icaps.v31i1.16007
|View full text |Cite
|
Sign up to set email alerts
|

Learning and Exploiting Shaped Reward Models for Large Scale Multiagent RL

Abstract: Many real world systems involve interaction among large number of agents to achieve a common goal, for example, air traffic control. Several model-free RL algorithms have been proposed for such settings. A key limitation is that the empirical reward signal in model-free case is not very effective in addressing the multiagent credit assignment problem, which determines an agent's contribution to the team's success. This results in lower solution quality and high sample complexity. To address this, we contribute… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 21 publications
0
1
0
Order By: Relevance
“…In order to achieve these objectives, each agent collaborates with neighboring nodes to increase its own reward and contribute to the rewards of others. Drawing inspiration from the principles of frequent measurability and spatial decomposability [57], we have defined a local reward function, r, for each intersection, which enables us to achieve these goals in a highly efficient and effective manner.…”
Section: ) Reward Functionmentioning
confidence: 99%
“…In order to achieve these objectives, each agent collaborates with neighboring nodes to increase its own reward and contribute to the rewards of others. Drawing inspiration from the principles of frequent measurability and spatial decomposability [57], we have defined a local reward function, r, for each intersection, which enables us to achieve these goals in a highly efficient and effective manner.…”
Section: ) Reward Functionmentioning
confidence: 99%