2016
DOI: 10.1177/0278364916679611
|View full text |Cite
|
Sign up to set email alerts
|

Policy search for multi-robot coordination under uncertainty

Abstract: We introduce a principled method for multi-robot coordination based on a general model (termed a MacDec-POMDP) of multi-robot cooperative planning in the presence of stochasticity, uncertain sensing, and communication limitations. A new MacDec-POMDP planning algorithm is presented that searches over policies represented as finite-state controllers, rather than the previous policy tree representation. Finite-state controllers can be much more concise than trees, are much easier to interpret, and can operate ove… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
43
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 46 publications
(43 citation statements)
references
References 26 publications
0
43
0
Order By: Relevance
“…Because tree-based representations become intractable as the horizon grows, we also developed multiple methods for optimizing finite-state controllers in macro-action Dec-POMDPs using the MacDec-POMDP [Amato et al, 2015a] and Dec-POSMDP [Omidshafiei et al, 2015] models. Some of these approaches can provide solutions with only a highlevel model of the macro-actions (i.e., distributions over time and outcomes) instead of a full model of the underlying Dec-POMDP [Amato et al, 2015a;2017;2017a]. Another approach automatically generates the macro-actions from low-level (continuous) dynamics models , while another method generates (macro-)observations from low-level sensor (e.g., camera) information [Omidshafiei et al, 2017b].…”
Section: Macro-action-based Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…Because tree-based representations become intractable as the horizon grows, we also developed multiple methods for optimizing finite-state controllers in macro-action Dec-POMDPs using the MacDec-POMDP [Amato et al, 2015a] and Dec-POSMDP [Omidshafiei et al, 2015] models. Some of these approaches can provide solutions with only a highlevel model of the macro-actions (i.e., distributions over time and outcomes) instead of a full model of the underlying Dec-POMDP [Amato et al, 2015a;2017;2017a]. Another approach automatically generates the macro-actions from low-level (continuous) dynamics models , while another method generates (macro-)observations from low-level sensor (e.g., camera) information [Omidshafiei et al, 2017b].…”
Section: Macro-action-based Methodsmentioning
confidence: 99%
“…We call this model a MacDec-POMDP [Amato et al, 2014] when the low-level Dec-POMDP model and the policies of the macro-actions are known and a decentralized partially observable semi-Markov decision process (Dec-POSMDP) when a high-level model is defined which includes time to completion Amato et al, 2015a] (but a simulator can be used in place of a model in each case). While these high-level models still include the states of the Dec-POMDP, they do not include the Dec-POMDP actions and observations.…”
Section: Macro-actions In Dec-pomdpsmentioning
confidence: 99%
See 2 more Smart Citations
“…Exploring algorithm determines the shortest and the lowest-cost path of a robot based on the environment size and type, cell size, and the numbers of robots constitute the exploration team and update the environment map based on the gathered information [3,24]. A set of common exploration algorithms will be discussed in section 4.…”
Section: Exploration Algorithmmentioning
confidence: 99%