2021
DOI: 10.4218/etrij.2021-0010
|View full text |Cite
|
Sign up to set email alerts
|

Avoiding collaborative paradox in multi‐agent reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(1 citation statement)
references
References 17 publications
0
1
0
Order By: Relevance
“…Liu et al proposed feudal latent space exploration for multi-agent reinforcement learning and guided a coordinated exploration using multiple agents by learning the latent structure [34]. Kim et al analyzed the problem of the collaboration paradox caused by "lazy" agents [35]. Kuba et al performed a rigorous mathematical analysis of the high variance in the estimations of the policy gradient method and derived the optimal baseline to achieve the minimum variance [36].…”
Section: A Related Workmentioning
confidence: 99%
“…Liu et al proposed feudal latent space exploration for multi-agent reinforcement learning and guided a coordinated exploration using multiple agents by learning the latent structure [34]. Kim et al analyzed the problem of the collaboration paradox caused by "lazy" agents [35]. Kuba et al performed a rigorous mathematical analysis of the high variance in the estimations of the policy gradient method and derived the optimal baseline to achieve the minimum variance [36].…”
Section: A Related Workmentioning
confidence: 99%