Proceedings of the Second International Joint Conference on Autonomous Agents and Multiagent Systems 2003
DOI: 10.1145/860575.860687
|View full text |Cite
|
Sign up to set email alerts
|

A selection-mutation model for q-learning in multi-agent systems

Abstract: Although well understood in the single-agent framework, the use of traditional reinforcement learning (RL) algorithms in multi-agent systems (MAS) is not always justified. The feedback an agent experiences in a MAS, is usually influenced by the other agents present in the system. Multi agent environments are therefore non-stationary and convergence and optimality guarantees of RL algorithms are lost. To better understand the dynamics of traditional RL algorithms we analyze the learning process in terms of evol… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
40
0

Year Published

2004
2004
2018
2018

Publication Types

Select...
5
1
1

Relationship

2
5

Authors

Journals

citations
Cited by 66 publications
(41 citation statements)
references
References 9 publications
(2 reference statements)
1
40
0
Order By: Relevance
“…In this section we briefly repeat the main results of [21]. In this paper however, we will extend the results of that previous work.…”
Section: The Q-learning Dynamicssupporting
confidence: 57%
See 2 more Smart Citations
“…In this section we briefly repeat the main results of [21]. In this paper however, we will extend the results of that previous work.…”
Section: The Q-learning Dynamicssupporting
confidence: 57%
“…More precisely, we will apply these results to stochastic Dispersion Games. The experiments of [21] were conducted in one-stage games. In this paper we will extend this approach to multi-state onestage games, i.e.…”
Section: The Q-learning Dynamicsmentioning
confidence: 99%
See 1 more Smart Citation
“…Many such tools are particularly suited to only certain learning methods, and only a few offer a common framework for multiple learning techniques. Evolutionary game theory is the main tool in this second category, and it was successfully used to study the properties of cooperative coevolution [78,296], to visualize basins of attraction to Nash equilibria for cooperative coevolution [191], and to study trajectories of concurrent Q-learning processes [271,263]. Another tool for modeling and predicting the dynamics of concurrent multi-agent learners was recently proposed in Vidal and Durfee [276,277].…”
Section: The Dynamics Of Learningmentioning
confidence: 99%
“…A selection-mutation model of Boltzmann Q-learning (Eqs. 1 and 2) has been proposed by Tuyls et al [36]. The dynamical system can again be decomposed into terms for exploitation (selection following the replicator dynamics) and exploration (mutation through randomization based on the Boltzmann mechanism):…”
Section: Evolutionary Models Of Multi-agent Learningmentioning
confidence: 99%