2010
DOI: 10.1007/978-3-642-11814-2_3
|View full text |Cite
|
Sign up to set email alerts
|

Replicator Dynamics for Multi-agent Learning: An Orthogonal Approach

Abstract: Today's society is largely connected and many real life applications lend themselves to be modeled as multi-agent systems. Although such systems as well as their models are desirable, e.g. for reasons of stability or parallelism, they are highly complex and therefore difficult to understand or predict. Multi-agent learning has been acknowledged to be indispensable to control or find solutions for such systems. Recently, evolutionary game theory has been linked to multi-agent reinforcement learning. However, ga… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 12 publications
0
4
0
Order By: Relevance
“…On a related note, and Tuyls et al (2006) considered certain classes of two-player matrix games and derived a connection between the exploration-exploitation scheme of RL and the selection-mutation mechanisms of evolutionary game theory; exploration being analogous to mutation and exploitation to selection. Further, there are links between RL and replicator dynamics; Kaisers and Tuyls (2010) showed empirical confirmation of the match between the learning trajectories of their frequency-adjusted Q-learning (FAQ-learning) algorithm and replicator dynamics for three 2 × 2 games, viz., Prisoner's Dilemma, Battle of Sexes, and Matching Pennies.…”
Section: Structuralmentioning
confidence: 80%
See 1 more Smart Citation
“…On a related note, and Tuyls et al (2006) considered certain classes of two-player matrix games and derived a connection between the exploration-exploitation scheme of RL and the selection-mutation mechanisms of evolutionary game theory; exploration being analogous to mutation and exploitation to selection. Further, there are links between RL and replicator dynamics; Kaisers and Tuyls (2010) showed empirical confirmation of the match between the learning trajectories of their frequency-adjusted Q-learning (FAQ-learning) algorithm and replicator dynamics for three 2 × 2 games, viz., Prisoner's Dilemma, Battle of Sexes, and Matching Pennies.…”
Section: Structuralmentioning
confidence: 80%
“…Despite these differences, existing research has found some connections between the three approaches. For example, the learning trajectories for RL and replicator dynamics were shown to match for certain PD games (Kaisers & Tuyls, 2010). However, many of the findings that connect the different learning approaches are not readily relevant to dilemma games.…”
Section: I5mentioning
confidence: 99%
“…The biggest advantage is that each agent/robot can implement its own RL algorithm and there is no need for coordination and, consequently, a joint policy calculation [16,27]. Independent classic Q-learners have shown promising results in AI [29,30], as well as in robotics [31,32]. On the other hand, the joint learners aim to learn the joint optimal policy from O and A.…”
Section: Multi-agent Q-learningmentioning
confidence: 99%
“…The more recent work of [2] explores a PAC (probably approximately correct) model for obtaining theoretical predictions for the value of coalitions that have not been observed in the past. The links between evolutionary game theory and multiagent reinforcement learning is the topic of study in [31,19].…”
Section: Uncertainty and Learningmentioning
confidence: 99%