2021
DOI: 10.48550/arxiv.2111.00781
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Decentralized Cooperative Reinforcement Learning with Hierarchical Information Structure

Abstract: Multi-agent reinforcement learning (MARL) problems are challenging due to information asymmetry. To overcome this challenge, existing methods often require high level of coordination or communication between the agents. We consider two-agent multi-armed bandits (MABs) and Markov decision processes (MDPs) with a hierarchical information structure arising in applications, which we exploit to propose simpler and more efficient algorithms that require no coordination or communication. In the structure, in each ste… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 21 publications
(17 reference statements)
0
1
0
Order By: Relevance
“…Another long line of work features collision models where rewards are lower if multiple agents simultaneously pull the same arm (e.g., [1,5,13,21,30,41,42,45,50]), unlike our model. Along these lines, other reward structures have been studied, such as reward being a function of the agents' joint action (e.g., [8,9,32]).…”
Section: Other Related Workmentioning
confidence: 99%
“…Another long line of work features collision models where rewards are lower if multiple agents simultaneously pull the same arm (e.g., [1,5,13,21,30,41,42,45,50]), unlike our model. Along these lines, other reward structures have been studied, such as reward being a function of the agents' joint action (e.g., [8,9,32]).…”
Section: Other Related Workmentioning
confidence: 99%