2022
DOI: 10.48550/arxiv.2204.08594
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Multi-UAV Collision Avoidance using Multi-Agent Reinforcement Learning with Counterfactual Credit Assignment

Abstract: Multi-UAV collision avoidance is a challenging task for UAV swarm applications due to the need of tight cooperation among swarm members for collision-free path planning. Centralized Training with Decentralized Execution (CTDE) in Multi-Agent Reinforcement Learning is a promising method for multi-UAV collision avoidance, in which the key challenge is to effectively learn decentralized policies that can maximize a global reward cooperatively. We propose a new multi-agent critic-actor learning scheme called MACA … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
1
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 14 publications
0
1
0
Order By: Relevance
“…In the lazy agent problem, the agents in a team are not performing equally well, but they are still receiving the same collective reward. To address this issue, researchers have suggested various learning and non-learning methods that assign credit to each agent based on their individual contributions [ 30 , 31 , 32 , 33 , 34 , 35 ]. Interestingly, the centralized training and decentralized execution indigenously have no issues or occurrences of lazy agents.…”
Section: Multi-agent Reinforcement Learningmentioning
confidence: 99%
“…In the lazy agent problem, the agents in a team are not performing equally well, but they are still receiving the same collective reward. To address this issue, researchers have suggested various learning and non-learning methods that assign credit to each agent based on their individual contributions [ 30 , 31 , 32 , 33 , 34 , 35 ]. Interestingly, the centralized training and decentralized execution indigenously have no issues or occurrences of lazy agents.…”
Section: Multi-agent Reinforcement Learningmentioning
confidence: 99%
“…The last column contains additional remarks, concerning advantages, limitations or other comments with regard to the relation between the mentioned methods and the approach described in this paper. Some of the proposed multi-agent systems are intended for solving a different task than collision avoidance, considered in this paper, such as formation control [22,23], training a team of USVs [24], searching water region by a team of UUV [25] and/or consider a different type of vehicle, such as UAV [26], UUV or a car [27].…”
Section: Multi-agent System For Usv Maneuver Auto-negotiationmentioning
confidence: 99%
“…The concept of the maneuver auto-negotiation system in relation to autonomous cars was proposed in [27], to UAV in [26] and to ships in [28][29][30]. In [28], collision avoidance and maneuver auto-negotiation is based on geometrical relationships.…”
Section: Multi-agent System For Usv Maneuver Auto-negotiationmentioning
confidence: 99%