2011 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning (ADPRL) 2011
DOI: 10.1109/adprl.2011.5967386
|View full text |Cite
|
Sign up to set email alerts
|

An adaptive-learning framework for semi-cooperative multi-agent coordination

Abstract: Abstract-Complex problems involving multiple agents exhibit varying degrees of cooperation. The levels of cooperation might reflect both differences in information as well as differences in goals. In this research, we develop a general mathematical model for distributed, semi-cooperative planning and suggest a solution strategy which involves decomposing the system into subproblems, each of which is specified at a certain period in time and controlled by an agent. The agents communicate marginal values of reso… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2013
2013
2022
2022

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(7 citation statements)
references
References 67 publications
(55 reference statements)
0
7
0
Order By: Relevance
“…It required 5s to compute the new gain (batch solution), and this time is used to fill the matrices Γ and δ of Eq. (15)(16), necessary for the calculation of Φ k i and y k i in Eq. ( 12).…”
Section: A Drone Model In Z-axis and Numerical Simulationmentioning
confidence: 99%
See 1 more Smart Citation
“…It required 5s to compute the new gain (batch solution), and this time is used to fill the matrices Γ and δ of Eq. (15)(16), necessary for the calculation of Φ k i and y k i in Eq. ( 12).…”
Section: A Drone Model In Z-axis and Numerical Simulationmentioning
confidence: 99%
“…A promising solution to solve the online control problem with plant uncertainties is the adaptive dynamic programming (ADP) method [14], [15]. This approach employs a 'forward-in-time' mechanism that looks for an optimal control policy by successively adapting two parametric structures, i.e., an action network and a critic network, to approximate the solution of the Hamilton-Jacobi-Bellman (HJB) equation [16], [17].…”
Section: Introductionmentioning
confidence: 99%
“…In the work of Boukhtouta et al, resource allocation techniques are utilized to activate and optimize the activities of agents. In the approach, to achieve the common objective of maximizing demand coverage of the system, agents can share information regarding the future value of resources, the demands of agents on resources, the local decisions of agents, and the corresponding contributions or impacts on resources to find the best decisions.…”
Section: Related Workmentioning
confidence: 99%
“…Our behavioral analytics framework is also related to research that explores stochastic control of multi-agent systems. Related methods include decentralized control (Li et al 2012), approximate dynamic programming (Boukhtouta et al 2011, George andPowell 2007), game-theoretic approaches (Adlakha and Johari 2013, Iyer et al 2011, Zhou et al 2016, and robust optimization (Blanchet et al 2013, Bertsimas and Goyal 2012, Lorca and Sun 2015. In general, these models consider very different settings from the ones we consider in this paper.…”
Section: Literature Reviewmentioning
confidence: 99%