2004
DOI: 10.1613/jair.1497
|View full text |Cite
|
Sign up to set email alerts
|

Solving Transition Independent Decentralized Markov Decision Processes

Abstract: Formal treatment of collaborative multi-agent systems has been lagging behind the rapid progress in sequential decision making by individual agents. Recent work in the area of decentralized Markov Decision Processes (MDPs) has contributed to closing this gap, but the computational complexity of these models remains a serious obstacle. To overcome this complexity barrier, we identify a specific class of decentralized MDPs in which the agents' transitions are independent. The class consists of independent colla… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

1
140
0

Year Published

2005
2005
2022
2022

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 116 publications
(141 citation statements)
references
References 20 publications
(28 reference statements)
1
140
0
Order By: Relevance
“…Some examples include distributed robot control [2,9] and networking problems [3,12]. In multiagent domains, robot sensors not only provide uncertain and incomplete information about their own state, but also about the location of the other robots.…”
mentioning
confidence: 99%
“…Some examples include distributed robot control [2,9] and networking problems [3,12]. In multiagent domains, robot sensors not only provide uncertain and incomplete information about their own state, but also about the location of the other robots.…”
mentioning
confidence: 99%
“…[18] proposes a solution based on Dynamic Programming (DP), while [19] extends point-based DP to the case of decentralized agents, and [20] applies heuristic search. Work in [21] has focused on a simpler model, denoted transition-independent where agents do not affect each other's state but cooperate via a joint reward signal instead. The difference between these algorithms and our work lies in their assumption that the agents' internal states are discrete and their modeling of a low number of action outcomes.…”
Section: Decision Theoretic Multiagent Planningmentioning
confidence: 99%
“…The assumption is realistic since many proprioceptive signals are accessible to the measure of the agent itself but are not communicated to other agents. Therefore the approach in [21] is of particular interest to us in this paper. Our model of interest is thus that of transition independent DECHMDPs (TI-DEC-HMDPs), where agents do not communicate but are subjected to internal resource constraints.…”
Section: Decision Theoretic Multiagent Planningmentioning
confidence: 99%
See 1 more Smart Citation
“…Recently, several researchers have concentrated on finding approximate solutions or finding new subsets of DEC-POMDP problems which are easier to solve and can model some real world problems (Becker et al 2004;Goldman and Zilberstein 2004). Although current approximate solution algorithms are able to solve slightly larger problems than exact solution algorithms, they still need considerable improvement in order to handle real world problems involving very large state spaces (Bernstein et al 2005;Nair et al 2003).…”
Section: Introductionmentioning
confidence: 99%