2011 Eighth International Conference on Fuzzy Systems and Knowledge Discovery (FSKD) 2011
DOI: 10.1109/fskd.2011.6019729
|View full text |Cite
|
Sign up to set email alerts
|

Multi-agent cooperation by reinforcement learning with teammate modeling and reward allotment

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0
1

Year Published

2019
2019
2022
2022

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 10 publications
0
6
0
1
Order By: Relevance
“…As an example, TM-LM-ASM (team-mate model-learning model-Action Selection Model) [30] is a JAL method that combines traditional Q-learning with a team-mate modeling mechanism. To do that, each learner has to memorize a If we examine the case of 4 agents which can make 5 actions in a 10 × 10 grid world, we obtain:…”
Section: Joint State / Joint Actionmentioning
confidence: 99%
See 2 more Smart Citations
“…As an example, TM-LM-ASM (team-mate model-learning model-Action Selection Model) [30] is a JAL method that combines traditional Q-learning with a team-mate modeling mechanism. To do that, each learner has to memorize a If we examine the case of 4 agents which can make 5 actions in a 10 × 10 grid world, we obtain:…”
Section: Joint State / Joint Actionmentioning
confidence: 99%
“…As explained earlier, the JAG method [18] ensures a global coordination while using independent learners but needs a centralized process to make sure that all agents choose the same joint action at each learning step, whereas the TM-LM-ASM method [30] is a full distributed learning approach that also provides a global coordination but employs joint action learners which make it unsuitable for systems considering many agents and/or large state spaces. Our objective is to develop a new intermediate approach between TM-LM-ASM and JAG.…”
Section: Proposed Reinforcement Learning Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…Every spot, visited by a vehicle, is marked to reduce the reward for visiting that spot again to prevent redundant work. In [18], estimation of teammates behavior for MARL is proposed for communication constrained scenarios by having each agent store a teammate model for all of its teammates and continuously update this model. Deep Learning is used for collision avoidance and path planning in [16] for noncommunicating agents.…”
Section: Introductionmentioning
confidence: 99%
“…Spychalski and Arendt proposed a methodology for implementing machine learning capability in multi-agent systems for aided design of selected control systems allowed to improve their performance by reducing the time spent processing requests that were previously acknowledged and stored in the learning module [12]. In [13], a new kind of multi-agent reinforcement learning algorithm, called TM_Qlearning, which combines traditional Q-learning with observation-based teammate modeling techniques, was proposed. Two multi-agent reinforcement learning methods, both consisting of promoting the selection of actions so that the chosen action not only relies on the present experience but also on an estimation of possible future ones, have been proposed to better solve the coordination problem and the exploration/exploitation dilemma in the case of nonstationary environments [14].…”
Section: Introductionmentioning
confidence: 99%